[ 515.308302] env[61936]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61936) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 515.308715] env[61936]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61936) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 515.308843] env[61936]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61936) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 515.309069] env[61936]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 515.402452] env[61936]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61936) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 515.412510] env[61936]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61936) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 516.011831] env[61936]: INFO nova.virt.driver [None req-e2bb3029-ca4b-4261-b23d-6ae084c05d66 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 516.084751] env[61936]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.084928] env[61936]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.085061] env[61936]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61936) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 519.387733] env[61936]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-f981fc9d-291b-4a4d-a3a2-fc7999c900fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.412774] env[61936]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61936) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 519.412976] env[61936]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-ba5488f2-69df-455c-b2b8-34ce8ff338ed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.445053] env[61936]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 0f19a. [ 519.445197] env[61936]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.360s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.445700] env[61936]: INFO nova.virt.vmwareapi.driver [None req-e2bb3029-ca4b-4261-b23d-6ae084c05d66 None None] VMware vCenter version: 7.0.3 [ 519.449201] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07af5ccf-e547-44b4-822a-7d5849c2016e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.466266] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccf28a1-0b1d-4bd5-beca-8d7dcb5bd809 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.471857] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a76e29-655f-4751-a7ea-bca66dee1e22 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.478263] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb40a27-8f09-4034-a6e6-b8494f94e3d3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.491044] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8fa14e-dc38-4412-9a0a-0b134f8532af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.496727] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd06f2b4-acf6-44f9-9e14-30f7b0ea51c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.527040] env[61936]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-b8931490-9297-4e5c-9e5e-5ed7db45d673 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.532182] env[61936]: DEBUG nova.virt.vmwareapi.driver [None req-e2bb3029-ca4b-4261-b23d-6ae084c05d66 None None] Extension org.openstack.compute already exists. {{(pid=61936) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 519.534861] env[61936]: INFO nova.compute.provider_config [None req-e2bb3029-ca4b-4261-b23d-6ae084c05d66 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 520.039057] env[61936]: DEBUG nova.context [None req-e2bb3029-ca4b-4261-b23d-6ae084c05d66 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),6c7a2b2b-d282-4d93-b09b-40c12dbce120(cell1) {{(pid=61936) load_cells /opt/stack/nova/nova/context.py:464}} [ 520.041423] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.041676] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.042356] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.042784] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Acquiring lock "6c7a2b2b-d282-4d93-b09b-40c12dbce120" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.042971] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Lock "6c7a2b2b-d282-4d93-b09b-40c12dbce120" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.043986] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Lock "6c7a2b2b-d282-4d93-b09b-40c12dbce120" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.065531] env[61936]: INFO dbcounter [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Registered counter for database nova_cell0 [ 520.073646] env[61936]: INFO dbcounter [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Registered counter for database nova_cell1 [ 520.076837] env[61936]: DEBUG oslo_db.sqlalchemy.engines [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61936) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 520.077217] env[61936]: DEBUG oslo_db.sqlalchemy.engines [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61936) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 520.081874] env[61936]: ERROR nova.db.main.api [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.081874] env[61936]: result = function(*args, **kwargs) [ 520.081874] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.081874] env[61936]: return func(*args, **kwargs) [ 520.081874] env[61936]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 520.081874] env[61936]: result = fn(*args, **kwargs) [ 520.081874] env[61936]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 520.081874] env[61936]: return f(*args, **kwargs) [ 520.081874] env[61936]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 520.081874] env[61936]: return db.service_get_minimum_version(context, binaries) [ 520.081874] env[61936]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 520.081874] env[61936]: _check_db_access() [ 520.081874] env[61936]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 520.081874] env[61936]: stacktrace = ''.join(traceback.format_stack()) [ 520.081874] env[61936]: [ 520.082849] env[61936]: ERROR nova.db.main.api [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.082849] env[61936]: result = function(*args, **kwargs) [ 520.082849] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.082849] env[61936]: return func(*args, **kwargs) [ 520.082849] env[61936]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 520.082849] env[61936]: result = fn(*args, **kwargs) [ 520.082849] env[61936]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 520.082849] env[61936]: return f(*args, **kwargs) [ 520.082849] env[61936]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 520.082849] env[61936]: return db.service_get_minimum_version(context, binaries) [ 520.082849] env[61936]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 520.082849] env[61936]: _check_db_access() [ 520.082849] env[61936]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 520.082849] env[61936]: stacktrace = ''.join(traceback.format_stack()) [ 520.082849] env[61936]: [ 520.083248] env[61936]: WARNING nova.objects.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Failed to get minimum service version for cell 6c7a2b2b-d282-4d93-b09b-40c12dbce120 [ 520.083372] env[61936]: WARNING nova.objects.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 520.083788] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Acquiring lock "singleton_lock" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.083947] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Acquired lock "singleton_lock" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.084205] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Releasing lock "singleton_lock" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.084524] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Full set of CONF: {{(pid=61936) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 520.084664] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ******************************************************************************** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 520.084791] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Configuration options gathered from: {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 520.084939] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 520.085153] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 520.085281] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ================================================================================ {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 520.085489] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] allow_resize_to_same_host = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.085654] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] arq_binding_timeout = 300 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.085783] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] backdoor_port = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.085907] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] backdoor_socket = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.086082] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] block_device_allocate_retries = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.086243] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] block_device_allocate_retries_interval = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.086407] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cert = self.pem {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.086568] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.086729] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute_monitors = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.086898] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] config_dir = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.087097] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] config_drive_format = iso9660 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.087233] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.087394] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] config_source = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.087556] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] console_host = devstack {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.087715] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] control_exchange = nova {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.087865] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cpu_allocation_ratio = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.088028] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] daemon = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.088195] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] debug = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.088345] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] default_access_ip_network_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.088504] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] default_availability_zone = nova {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.088652] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] default_ephemeral_format = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.088803] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] default_green_pool_size = 1000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.089039] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.089203] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] default_schedule_zone = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.089375] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] disk_allocation_ratio = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.089532] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] enable_new_services = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.089701] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] enabled_apis = ['osapi_compute'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.089857] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] enabled_ssl_apis = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.090014] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] flat_injected = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.090170] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] force_config_drive = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.090324] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] force_raw_images = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.090486] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] graceful_shutdown_timeout = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.090638] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] heal_instance_info_cache_interval = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.090838] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] host = cpu-1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.091015] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.091184] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.091337] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.091544] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.091701] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instance_build_timeout = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.091854] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instance_delete_interval = 300 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.092017] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instance_format = [instance: %(uuid)s] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.092482] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instance_name_template = instance-%08x {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.092482] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instance_usage_audit = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.092482] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instance_usage_audit_period = month {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.092628] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.092786] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.092945] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] internal_service_availability_zone = internal {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.093108] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] key = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.093263] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] live_migration_retry_count = 30 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.093426] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_color = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.093584] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_config_append = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.093747] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.093918] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_dir = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.094131] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.094260] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_options = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.094420] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_rotate_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.094582] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_rotate_interval_type = days {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.094740] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] log_rotation_type = none {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.094864] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.094988] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.095163] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.095325] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.095447] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.095601] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] long_rpc_timeout = 1800 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.095752] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] max_concurrent_builds = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.095900] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] max_concurrent_live_migrations = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.096061] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] max_concurrent_snapshots = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.096219] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] max_local_block_devices = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.096370] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] max_logfile_count = 30 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.096519] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] max_logfile_size_mb = 200 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.096669] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] maximum_instance_delete_attempts = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.096832] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] metadata_listen = 0.0.0.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.097031] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] metadata_listen_port = 8775 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.097208] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] metadata_workers = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.097367] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] migrate_max_retries = -1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.097532] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] mkisofs_cmd = genisoimage {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.097732] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.097861] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] my_ip = 10.180.1.21 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.098069] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.098230] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] network_allocate_retries = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.098401] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.098563] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.098719] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] osapi_compute_listen_port = 8774 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.098881] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] osapi_compute_unique_server_name_scope = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.099056] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] osapi_compute_workers = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.099218] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] password_length = 12 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.099372] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] periodic_enable = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.099523] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] periodic_fuzzy_delay = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.099683] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] pointer_model = usbtablet {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.099842] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] preallocate_images = none {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.099994] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] publish_errors = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.100131] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] pybasedir = /opt/stack/nova {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.100283] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ram_allocation_ratio = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.100434] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] rate_limit_burst = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.100593] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] rate_limit_except_level = CRITICAL {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.100742] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] rate_limit_interval = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.100892] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] reboot_timeout = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.101059] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] reclaim_instance_interval = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.101212] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] record = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.101373] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] reimage_timeout_per_gb = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.101535] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] report_interval = 120 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.101691] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] rescue_timeout = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.101843] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] reserved_host_cpus = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.101995] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] reserved_host_disk_mb = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.102165] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] reserved_host_memory_mb = 512 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.102318] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] reserved_huge_pages = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.102468] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] resize_confirm_window = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.102622] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] resize_fs_using_block_device = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.102772] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] resume_guests_state_on_host_boot = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.102930] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.103095] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] rpc_response_timeout = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.103251] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] run_external_periodic_tasks = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.103412] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] running_deleted_instance_action = reap {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.103564] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.103713] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] running_deleted_instance_timeout = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.103864] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler_instance_sync_interval = 120 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.104034] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_down_time = 720 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.104200] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] servicegroup_driver = db {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.104350] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] shell_completion = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.104502] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] shelved_offload_time = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.104651] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] shelved_poll_interval = 3600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.104808] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] shutdown_timeout = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.104978] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] source_is_ipv6 = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.105159] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ssl_only = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.105399] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.105560] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] sync_power_state_interval = 600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.105714] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] sync_power_state_pool_size = 1000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.105875] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] syslog_log_facility = LOG_USER {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.106043] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] tempdir = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.106203] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] timeout_nbd = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.106361] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] transport_url = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.106514] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] update_resources_interval = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.106667] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] use_cow_images = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.106817] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] use_eventlog = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.106997] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] use_journal = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.107176] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] use_json = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.107332] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] use_rootwrap_daemon = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.107489] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] use_stderr = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.107643] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] use_syslog = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.107793] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vcpu_pin_set = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.108045] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plugging_is_fatal = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.108238] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plugging_timeout = 300 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.108404] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] virt_mkfs = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.108562] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] volume_usage_poll_interval = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.108720] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] watch_log_file = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.108882] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] web = /usr/share/spice-html5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 520.109074] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.109244] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.109400] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.109564] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_concurrency.disable_process_locking = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.109833] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.110014] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.110187] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.110354] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.110519] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.110678] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.110850] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.auth_strategy = keystone {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.111012] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.compute_link_prefix = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.111186] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.111355] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.dhcp_domain = novalocal {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.111514] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.enable_instance_password = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.111670] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.glance_link_prefix = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.111826] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.111989] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.112169] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.instance_list_per_project_cells = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.112324] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.list_records_by_skipping_down_cells = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.112478] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.local_metadata_per_cell = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.112637] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.max_limit = 1000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.112798] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.metadata_cache_expiration = 15 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.112964] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.neutron_default_tenant_id = default {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.113145] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.response_validation = warn {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.113307] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.use_neutron_default_nets = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.113472] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.113627] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.113785] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.113948] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.114125] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.vendordata_dynamic_targets = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.114284] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.vendordata_jsonfile_path = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.114455] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.114638] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.backend = dogpile.cache.memcached {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.114797] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.backend_argument = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.114978] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.config_prefix = cache.oslo {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.115168] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.dead_timeout = 60.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.115328] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.debug_cache_backend = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.115481] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.enable_retry_client = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.115634] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.enable_socket_keepalive = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.115798] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.enabled = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.115955] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.enforce_fips_mode = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.116127] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.expiration_time = 600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.116284] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.hashclient_retry_attempts = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.116441] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.116596] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_dead_retry = 300 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.116750] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_password = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.116931] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.117115] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.117281] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_pool_maxsize = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.117440] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.117595] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_sasl_enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.117769] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.117934] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.118101] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.memcache_username = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.118264] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.proxies = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.118422] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.redis_db = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.118574] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.redis_password = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.118739] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.118909] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.119083] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.redis_server = localhost:6379 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.119261] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.redis_socket_timeout = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.119427] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.redis_username = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.119584] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.retry_attempts = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.119742] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.retry_delay = 0.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.119898] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.socket_keepalive_count = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.120066] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.socket_keepalive_idle = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.120223] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.socket_keepalive_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.120375] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.tls_allowed_ciphers = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.120525] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.tls_cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.120673] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.tls_certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.120826] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.tls_enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.120975] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cache.tls_keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.121153] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.121322] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.auth_type = password {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.121478] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.121644] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.121799] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.121958] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.122129] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.cross_az_attach = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.122287] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.debug = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.122439] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.endpoint_template = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.122599] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.http_retries = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.122756] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.122908] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.123087] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.os_region_name = RegionOne {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.123247] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.123401] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cinder.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.123651] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.123843] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.cpu_dedicated_set = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.124010] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.cpu_shared_set = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.124186] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.image_type_exclude_list = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.124348] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.124507] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.124664] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.124824] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.125023] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.125200] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.resource_provider_association_refresh = 300 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.125372] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.125531] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.shutdown_retry_interval = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.125706] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.125877] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] conductor.workers = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.126061] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] console.allowed_origins = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.126224] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] console.ssl_ciphers = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.126388] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] console.ssl_minimum_version = default {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.126551] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] consoleauth.enforce_session_timeout = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.126713] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] consoleauth.token_ttl = 600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.126879] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.127078] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.127248] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.127401] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.127553] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.127702] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.127859] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.128047] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.128212] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.128417] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.128514] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.region_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.128666] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.128817] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.128980] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.service_type = accelerator {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.129149] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.129306] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.129458] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.129612] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.129788] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.129947] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] cyborg.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.130137] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.backend = sqlalchemy {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.130307] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.connection = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.130468] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.connection_debug = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.130633] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.connection_parameters = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.130793] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.connection_recycle_time = 3600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.130953] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.connection_trace = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.131125] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.db_inc_retry_interval = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.131284] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.db_max_retries = 20 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.131441] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.db_max_retry_interval = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.131596] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.db_retry_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.131751] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.max_overflow = 50 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.131908] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.max_pool_size = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.132076] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.max_retries = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.132248] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.132398] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.mysql_wsrep_sync_wait = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.132548] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.pool_timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.132702] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.retry_interval = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.132855] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.slave_connection = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.133016] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.sqlite_synchronous = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.133176] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] database.use_db_reconnect = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.133350] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.backend = sqlalchemy {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.133516] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.connection = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.133673] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.connection_debug = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.133835] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.connection_parameters = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.133994] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.connection_recycle_time = 3600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.134166] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.connection_trace = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.134323] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.db_inc_retry_interval = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.134480] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.db_max_retries = 20 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.134636] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.db_max_retry_interval = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.134794] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.db_retry_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.134971] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.max_overflow = 50 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.135157] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.max_pool_size = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.135318] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.max_retries = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.135486] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.135644] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.135795] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.pool_timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.135951] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.retry_interval = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.136119] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.slave_connection = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.136281] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] api_database.sqlite_synchronous = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.136454] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] devices.enabled_mdev_types = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.136629] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.136794] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.136974] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ephemeral_storage_encryption.enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.137157] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.137332] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.api_servers = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.137492] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.137649] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.137808] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.137964] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.138134] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.138294] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.debug = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.138455] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.default_trusted_certificate_ids = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.138611] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.enable_certificate_validation = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.138769] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.enable_rbd_download = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.138924] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.139102] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.139265] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.139419] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.139571] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.139728] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.num_retries = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.139894] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.rbd_ceph_conf = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.140063] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.rbd_connect_timeout = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.140231] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.rbd_pool = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.140394] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.rbd_user = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.140551] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.region_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.140705] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.140858] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.141028] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.service_type = image {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.141201] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.141353] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.141504] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.141658] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.141832] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.141991] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.verify_glance_signatures = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.142219] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] glance.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.142409] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] guestfs.debug = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.142578] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.142741] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.auth_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.142897] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.143067] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.143232] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.143388] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.143542] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.143695] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.143853] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.144013] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.144179] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.144331] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.144484] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.region_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.144637] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.144789] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.144972] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.service_type = shared-file-system {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.145161] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.share_apply_policy_timeout = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.145327] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.145480] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.145635] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.145789] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.145966] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.146141] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] manila.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.146310] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] mks.enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.146665] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.146855] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] image_cache.manager_interval = 2400 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.147053] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] image_cache.precache_concurrency = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.147232] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] image_cache.remove_unused_base_images = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.147400] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.147567] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.147738] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] image_cache.subdirectory_name = _base {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.147910] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.api_max_retries = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.148094] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.api_retry_interval = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.148256] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.148415] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.auth_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.148572] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.148726] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.148884] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.149053] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.conductor_group = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.149215] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.149370] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.149522] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.149682] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.149836] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.149988] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.150156] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.150318] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.peer_list = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.150473] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.region_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.150625] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.150785] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.serial_console_state_timeout = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.150942] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.151124] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.service_type = baremetal {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.151281] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.shard = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.151440] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.151596] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.151751] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.151905] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.152101] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.152264] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ironic.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.152441] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.152611] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] key_manager.fixed_key = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.152790] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.152950] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.barbican_api_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.153123] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.barbican_endpoint = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.153290] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.barbican_endpoint_type = public {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.153446] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.barbican_region_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.153599] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.153752] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.153913] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.154083] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.154243] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.154401] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.number_of_retries = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.154558] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.retry_delay = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.154717] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.send_service_user_token = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.154876] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.155068] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.155239] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.verify_ssl = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.155396] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican.verify_ssl_path = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.155557] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.155717] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.auth_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.155871] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.156032] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.156200] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.156359] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.156512] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.156670] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.156825] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] barbican_service_user.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.157020] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.approle_role_id = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.157190] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.approle_secret_id = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.157357] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.kv_mountpoint = secret {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.157516] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.kv_path = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.157677] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.kv_version = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.157835] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.namespace = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.158023] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.root_token_id = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.158183] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.ssl_ca_crt_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.158354] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.timeout = 60.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.158512] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.use_ssl = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.158675] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.158839] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.158997] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.auth_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.159170] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.159324] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.159483] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.159638] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.159792] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.159945] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.160115] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.160270] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.160424] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.160573] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.160726] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.region_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.160879] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.161041] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.161215] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.service_type = identity {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.161377] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.161535] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.161692] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.161845] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.162028] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.162192] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] keystone.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.162379] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.ceph_mount_options = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.163047] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.163244] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.connection_uri = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.163413] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.cpu_mode = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.163583] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.163753] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.cpu_models = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.163924] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.cpu_power_governor_high = performance {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.164108] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.164276] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.cpu_power_management = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.164445] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.164614] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.device_detach_attempts = 8 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.164771] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.device_detach_timeout = 20 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.164962] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.disk_cachemodes = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.165162] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.disk_prefix = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.165338] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.enabled_perf_events = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.165501] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.file_backed_memory = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.165666] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.gid_maps = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.165822] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.hw_disk_discard = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.165977] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.hw_machine_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.166163] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.images_rbd_ceph_conf = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.166326] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.166485] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.166648] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.images_rbd_glance_store_name = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.166814] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.images_rbd_pool = rbd {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.167008] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.images_type = default {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.167183] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.images_volume_group = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.167345] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.inject_key = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.167504] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.inject_partition = -2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.167663] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.inject_password = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.167823] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.iscsi_iface = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.167982] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.iser_use_multipath = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.168161] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.168321] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.168481] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_downtime = 500 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.168643] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.168802] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.168958] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_inbound_addr = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.169132] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.169293] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.169445] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_scheme = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.169614] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_timeout_action = abort {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.169774] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_tunnelled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.169932] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_uri = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.170106] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.live_migration_with_native_tls = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.170268] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.max_queues = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.170428] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.170659] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.170824] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.nfs_mount_options = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.171114] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.171293] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.171457] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.171618] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.171781] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.171945] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.num_pcie_ports = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.172131] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.172302] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.pmem_namespaces = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.172462] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.quobyte_client_cfg = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.172745] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.172916] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.173094] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.173258] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.173418] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rbd_secret_uuid = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.173574] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rbd_user = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.173733] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.173901] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.174072] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rescue_image_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.174235] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rescue_kernel_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.174391] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rescue_ramdisk_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.174558] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.174712] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.rx_queue_size = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.174877] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.smbfs_mount_options = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.175187] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.175363] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.snapshot_compression = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.175523] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.snapshot_image_format = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.175740] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.175904] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.sparse_logical_volumes = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.176079] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.swtpm_enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.176250] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.swtpm_group = tss {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.176418] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.swtpm_user = tss {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.176583] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.sysinfo_serial = unique {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.176740] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.tb_cache_size = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.176905] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.tx_queue_size = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.177096] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.uid_maps = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.177265] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.use_virtio_for_bridges = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.177435] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.virt_type = kvm {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.177602] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.volume_clear = zero {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.177763] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.volume_clear_size = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.177927] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.volume_use_multipath = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.178122] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.vzstorage_cache_path = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.178303] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.178472] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.178637] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.178804] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.179090] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.179273] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.vzstorage_mount_user = stack {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.179437] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.179608] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.179778] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.auth_type = password {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.179937] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.180105] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.180269] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.180425] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.180582] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.180747] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.default_floating_pool = public {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.180903] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.181095] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.extension_sync_interval = 600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.181273] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.http_retries = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.181435] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.181594] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.181748] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.181913] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.182088] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.182258] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.ovs_bridge = br-int {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.182421] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.physnets = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.182586] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.region_name = RegionOne {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.182741] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.182907] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.service_metadata_proxy = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.183077] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.183246] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.service_type = network {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.183408] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.183562] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.183718] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.183874] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.184072] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.184256] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] neutron.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.184429] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] notifications.bdms_in_notifications = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.184606] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] notifications.default_level = INFO {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.184779] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] notifications.notification_format = unversioned {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.184960] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] notifications.notify_on_state_change = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.185163] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.185342] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] pci.alias = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.185509] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] pci.device_spec = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.185670] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] pci.report_in_placement = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.185837] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.186021] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.auth_type = password {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.186193] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.186352] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.186507] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.186668] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.186821] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.187025] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.187186] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.default_domain_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.187342] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.default_domain_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.187496] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.domain_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.187651] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.domain_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.187808] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.187965] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.188134] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.188290] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.188443] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.188607] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.password = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.188762] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.project_domain_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.188926] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.project_domain_name = Default {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.189103] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.project_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.189278] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.project_name = service {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.189442] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.region_name = RegionOne {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.189603] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.189758] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.189922] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.service_type = placement {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.190097] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.190257] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.190414] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.190569] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.system_scope = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.190725] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.190880] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.trust_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.191045] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.user_domain_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.191216] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.user_domain_name = Default {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.191373] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.user_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.191544] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.username = nova {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.191721] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.191877] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] placement.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.192070] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.cores = 20 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.192243] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.count_usage_from_placement = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.192412] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.192584] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.injected_file_content_bytes = 10240 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.192747] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.injected_file_path_length = 255 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.192907] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.injected_files = 5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.193081] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.instances = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.193247] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.key_pairs = 100 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.193408] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.metadata_items = 128 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.193568] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.ram = 51200 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.193727] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.recheck_quota = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.193892] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.server_group_members = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.194066] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] quota.server_groups = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.194240] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.194403] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.194563] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.image_metadata_prefilter = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.194723] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.194896] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.max_attempts = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.195083] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.max_placement_results = 1000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.195254] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.195416] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.195574] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.195743] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] scheduler.workers = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.195911] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.196088] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.196268] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.196431] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.196591] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.196753] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.196931] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.197128] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.197299] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.host_subset_size = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.197464] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.197624] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.197783] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.197943] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.isolated_hosts = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.198117] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.isolated_images = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.198278] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.198437] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.198600] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.198760] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.pci_in_placement = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.198920] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.199107] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.199272] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.199427] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.199586] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.199743] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.199900] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.track_instance_changes = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.200086] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.200256] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] metrics.required = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.200417] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] metrics.weight_multiplier = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.200579] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.200740] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] metrics.weight_setting = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.201058] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.201235] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] serial_console.enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.201409] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] serial_console.port_range = 10000:20000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.201576] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.201742] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.201905] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] serial_console.serialproxy_port = 6083 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.202082] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.202258] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.auth_type = password {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.202414] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.202570] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.202730] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.202886] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.203049] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.203221] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.send_service_user_token = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.203382] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.203537] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] service_user.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.203701] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.agent_enabled = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.203860] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.204168] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.204357] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.204522] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.html5proxy_port = 6082 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.204679] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.image_compression = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.204847] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.jpeg_compression = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.205029] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.playback_compression = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.205204] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.require_secure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.205369] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.server_listen = 127.0.0.1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.205533] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.205688] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.streaming_mode = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.205841] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] spice.zlib_compression = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.206011] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] upgrade_levels.baseapi = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.206185] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] upgrade_levels.compute = auto {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.206346] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] upgrade_levels.conductor = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.206501] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] upgrade_levels.scheduler = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.206662] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.206818] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.207012] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.207178] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.207399] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.207553] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.207768] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.207977] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.208199] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vendordata_dynamic_auth.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.208397] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.api_retry_count = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.208560] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.ca_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.208731] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.208896] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.cluster_name = testcl1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.209153] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.connection_pool_size = 10 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.209322] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.console_delay_seconds = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.209490] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.datastore_regex = ^datastore.* {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.209695] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.209866] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.host_password = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.210041] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.host_port = 443 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.210235] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.host_username = administrator@vsphere.local {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.210406] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.insecure = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.210567] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.integration_bridge = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.210726] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.maximum_objects = 100 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.210882] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.pbm_default_policy = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.211051] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.pbm_enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.211213] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.pbm_wsdl_location = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.211378] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.211534] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.serial_port_proxy_uri = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.211687] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.serial_port_service_uri = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.211848] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.task_poll_interval = 0.5 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.212024] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.use_linked_clone = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.212197] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.vnc_keymap = en-us {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.212359] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.vnc_port = 5900 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.212516] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vmware.vnc_port_total = 10000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.212700] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.auth_schemes = ['none'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.212873] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.213177] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.213363] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.213529] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.novncproxy_port = 6080 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.213704] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.server_listen = 127.0.0.1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.213872] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.214042] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.vencrypt_ca_certs = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.214205] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.vencrypt_client_cert = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.214360] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vnc.vencrypt_client_key = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.214536] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.214693] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.disable_deep_image_inspection = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.214849] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.215013] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.215183] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.215340] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.disable_rootwrap = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.215497] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.enable_numa_live_migration = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.215653] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.215808] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.215963] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.216133] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.libvirt_disable_apic = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.216288] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.216445] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.216599] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.216753] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.216932] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.217107] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.217273] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.217432] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.217590] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.217751] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.217929] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.218130] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.client_socket_timeout = 900 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.218304] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.default_pool_size = 1000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.218469] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.keep_alive = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.218658] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.max_header_line = 16384 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.218820] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.218979] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.ssl_ca_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.219149] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.ssl_cert_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.219333] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.ssl_key_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.219504] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.tcp_keepidle = 600 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.219679] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.219846] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] zvm.ca_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.220010] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] zvm.cloud_connector_url = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.220306] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.220487] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] zvm.reachable_timeout = 300 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.220654] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.enforce_new_defaults = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.221036] env[61936]: WARNING oslo_config.cfg [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 520.221223] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.enforce_scope = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.221397] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.policy_default_rule = default {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.221575] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.221748] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.policy_file = policy.yaml {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.221918] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.222094] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.222256] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.222411] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.222568] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.222732] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.222906] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.223094] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.connection_string = messaging:// {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.223263] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.enabled = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.223431] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.es_doc_type = notification {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.223594] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.es_scroll_size = 10000 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.223760] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.es_scroll_time = 2m {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.223920] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.filter_error_trace = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.224101] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.hmac_keys = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.224268] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.sentinel_service_name = mymaster {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.224429] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.socket_timeout = 0.1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.224587] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.trace_requests = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.224745] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler.trace_sqlalchemy = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.224924] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler_jaeger.process_tags = {} {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.225095] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler_jaeger.service_name_prefix = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.225266] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] profiler_otlp.service_name_prefix = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.225440] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] remote_debug.host = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.225598] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] remote_debug.port = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.225777] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.225937] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.226113] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.226274] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.226432] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.226587] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.226748] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.226925] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.227100] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.227273] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.227431] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.227600] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.227762] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.227930] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.228109] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.228276] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.228434] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.228604] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.228764] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.228923] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.229102] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.229284] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.229457] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.229625] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.229786] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.229946] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.230120] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.230282] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.230445] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.230608] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.ssl = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.230778] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.230943] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.231118] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.231287] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.231454] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.231613] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.231793] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.231955] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_notifications.retry = -1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.232148] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.232320] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.232486] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.auth_section = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.232644] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.auth_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.232797] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.cafile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.232949] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.certfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.233122] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.collect_timing = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.233278] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.connect_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.233433] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.connect_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.233585] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.endpoint_id = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.233737] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.endpoint_override = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.233895] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.insecure = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.234065] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.keyfile = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.234223] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.max_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.234375] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.min_version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.234528] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.region_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.234685] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.retriable_status_codes = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.234842] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.service_name = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.234996] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.service_type = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.235169] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.split_loggers = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.235324] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.status_code_retries = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.235477] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.status_code_retry_delay = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.235629] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.timeout = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.235782] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.valid_interfaces = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.235934] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_limit.version = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.236107] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_reports.file_event_handler = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.236266] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.236419] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] oslo_reports.log_dir = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.236582] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.236735] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.236890] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.237087] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.237254] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.237409] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.237571] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.237723] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_ovs_privileged.group = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.237875] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.238069] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.238233] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.238388] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] vif_plug_ovs_privileged.user = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.238551] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.238722] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.238888] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.239069] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.239261] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.239513] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.239750] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.239996] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.240277] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.240533] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_ovs.isolate_vif = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.240807] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.241092] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.241351] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.241542] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.241709] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] os_vif_ovs.per_port_bridge = False {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.241878] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] privsep_osbrick.capabilities = [21] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.242048] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] privsep_osbrick.group = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.242211] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] privsep_osbrick.helper_command = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.242374] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.242533] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.242686] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] privsep_osbrick.user = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.242854] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.243014] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] nova_sys_admin.group = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.243177] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] nova_sys_admin.helper_command = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.243338] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.243494] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.243646] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] nova_sys_admin.user = None {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 520.243774] env[61936]: DEBUG oslo_service.service [None req-b5f9228b-8f7f-40a8-a830-cd7798aef914 None None] ******************************************************************************** {{(pid=61936) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 520.244266] env[61936]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 520.747382] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Getting list of instances from cluster (obj){ [ 520.747382] env[61936]: value = "domain-c8" [ 520.747382] env[61936]: _type = "ClusterComputeResource" [ 520.747382] env[61936]: } {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 520.748544] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdfb4d8-6f79-489d-b2ed-3732f89f1a0a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.757601] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Got total of 0 instances {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 520.758162] env[61936]: WARNING nova.virt.vmwareapi.driver [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 520.758656] env[61936]: INFO nova.virt.node [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Generated node identity abd382a7-b427-4dcc-a050-3fb56fdd60ac [ 520.758889] env[61936]: INFO nova.virt.node [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Wrote node identity abd382a7-b427-4dcc-a050-3fb56fdd60ac to /opt/stack/data/n-cpu-1/compute_id [ 521.261845] env[61936]: WARNING nova.compute.manager [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Compute nodes ['abd382a7-b427-4dcc-a050-3fb56fdd60ac'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 522.267538] env[61936]: INFO nova.compute.manager [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 523.273515] env[61936]: WARNING nova.compute.manager [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 523.273885] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.274014] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.274142] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.274300] env[61936]: DEBUG nova.compute.resource_tracker [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 523.275260] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd5f27a-f1f0-47a2-987f-f4b02bd9c1f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.283773] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4766940b-e13e-41e7-b7f9-553f1dfc7cfa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.299108] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bfeef2-ec14-4335-b2f5-5bff42c96ad4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.305120] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10da727a-d9d8-43ab-a366-2e7af88342d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.333362] env[61936]: DEBUG nova.compute.resource_tracker [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181521MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 523.333533] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.333741] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.836087] env[61936]: WARNING nova.compute.resource_tracker [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] No compute node record for cpu-1:abd382a7-b427-4dcc-a050-3fb56fdd60ac: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host abd382a7-b427-4dcc-a050-3fb56fdd60ac could not be found. [ 524.340422] env[61936]: INFO nova.compute.resource_tracker [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: abd382a7-b427-4dcc-a050-3fb56fdd60ac [ 525.848927] env[61936]: DEBUG nova.compute.resource_tracker [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 525.849398] env[61936]: DEBUG nova.compute.resource_tracker [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 526.005140] env[61936]: INFO nova.scheduler.client.report [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] [req-5758ccea-827b-4bf8-b48b-14c6c20bb346] Created resource provider record via placement API for resource provider with UUID abd382a7-b427-4dcc-a050-3fb56fdd60ac and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 526.020993] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eddb9fd-2552-4086-9979-23836e523553 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.028705] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edae9c77-323d-4f31-ac67-02d035674097 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.057766] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cb9deb-c596-4254-9cba-38fb5805880b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.064583] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6668c2c0-68a0-4e88-929a-3045f38b156a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.077164] env[61936]: DEBUG nova.compute.provider_tree [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 526.614523] env[61936]: DEBUG nova.scheduler.client.report [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 526.614752] env[61936]: DEBUG nova.compute.provider_tree [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 0 to 1 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 526.614929] env[61936]: DEBUG nova.compute.provider_tree [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 526.663788] env[61936]: DEBUG nova.compute.provider_tree [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 1 to 2 during operation: update_traits {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 527.168183] env[61936]: DEBUG nova.compute.resource_tracker [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 527.168541] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.835s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.168541] env[61936]: DEBUG nova.service [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Creating RPC server for service compute {{(pid=61936) start /opt/stack/nova/nova/service.py:186}} [ 527.183572] env[61936]: DEBUG nova.service [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] Join ServiceGroup membership for this service compute {{(pid=61936) start /opt/stack/nova/nova/service.py:203}} [ 527.183747] env[61936]: DEBUG nova.servicegroup.drivers.db [None req-6ee2e08c-e6eb-4131-b968-c35c8cc4e4e4 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61936) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 558.185721] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_power_states {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.689250] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Getting list of instances from cluster (obj){ [ 558.689250] env[61936]: value = "domain-c8" [ 558.689250] env[61936]: _type = "ClusterComputeResource" [ 558.689250] env[61936]: } {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 558.690543] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f239eb45-efe3-4ebd-88cd-70a136c4f443 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.700483] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Got total of 0 instances {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 558.700483] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 558.700483] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Getting list of instances from cluster (obj){ [ 558.700483] env[61936]: value = "domain-c8" [ 558.700483] env[61936]: _type = "ClusterComputeResource" [ 558.700483] env[61936]: } {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 558.701451] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22162d48-01f8-41d7-b5d3-ca15b16b9062 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.709116] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Got total of 0 instances {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 563.007426] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Acquiring lock "bc67f44d-e6c8-49e0-ac2a-cf7449253580" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.008112] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Lock "bc67f44d-e6c8-49e0-ac2a-cf7449253580" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.511210] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 564.050393] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.050689] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.052851] env[61936]: INFO nova.compute.claims [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.836492] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "c4b372a3-f76a-411d-a336-338f5c6d502c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.836741] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "c4b372a3-f76a-411d-a336-338f5c6d502c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.123364] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d64d67-f35b-4ae5-bb57-b2550ec2a24a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.132247] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a0d51c-31ea-4dc8-912d-d2c0e1d36409 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.167799] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f1ea93-aba5-4ebd-9c56-3716510b97d7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.177622] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d671bddf-84ff-4e0e-8960-809e7c84b42b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.194257] env[61936]: DEBUG nova.compute.provider_tree [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.340016] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Acquiring lock "449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.340016] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Lock "449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.342275] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 565.700079] env[61936]: DEBUG nova.scheduler.client.report [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 565.845469] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 565.880868] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.059406] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "7dadef86-7685-40e2-9e68-debc4b4a8a95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.059578] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "7dadef86-7685-40e2-9e68-debc4b4a8a95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.134772] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Acquiring lock "3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.134772] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Lock "3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.205462] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.155s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.206235] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 566.212034] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.328s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.212034] env[61936]: INFO nova.compute.claims [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.387405] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.509974] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Acquiring lock "5aa38a34-2f1c-4546-9688-980d65dfed27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.509974] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Lock "5aa38a34-2f1c-4546-9688-980d65dfed27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.564041] env[61936]: DEBUG nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 566.637459] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 566.720339] env[61936]: DEBUG nova.compute.utils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 566.720339] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 566.720339] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 567.013781] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 567.098460] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.121214] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Acquiring lock "ffdb44ea-791a-41b3-9e65-9cf77717ce9a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.121214] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Lock "ffdb44ea-791a-41b3-9e65-9cf77717ce9a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.164550] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.228116] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 567.371615] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea59909-6f26-4435-aa19-6f66f503b518 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.380448] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddd4282-d8f9-4b46-9e95-14c27ca39021 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.417264] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf169b5-2345-452e-aa6b-c44d0fa97cf3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.427725] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed61b70c-5ddc-4ba9-9792-65371aeaf9f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.443815] env[61936]: DEBUG nova.compute.provider_tree [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.552633] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.626592] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 567.755023] env[61936]: DEBUG nova.policy [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf5b371d5cc04f20b94dca773127dff5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5c3d58b646c846f3ad952b6ef323b07c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.948602] env[61936]: DEBUG nova.scheduler.client.report [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 568.161853] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.209256] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "09467d10-8235-4712-b00f-d74bc6ed39cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.209555] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "09467d10-8235-4712-b00f-d74bc6ed39cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.240442] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 568.281755] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 568.283210] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.283210] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 568.283210] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.283210] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 568.283431] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 568.283780] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 568.284557] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 568.285537] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 568.285831] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 568.286303] env[61936]: DEBUG nova.virt.hardware [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 568.288113] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6ce961-be21-499d-b8a3-c4c23cb19522 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.297680] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c515c4b3-9bc0-403e-b924-3c474eda76b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.314253] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2041be4f-8235-4c88-8a74-f0f55d663dbe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.350731] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Successfully created port: 37e2f9e8-3071-4377-95ab-e032be86f90e {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.456498] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.456677] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 568.459492] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.072s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.461405] env[61936]: INFO nova.compute.claims [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.713698] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 568.967886] env[61936]: DEBUG nova.compute.utils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.976097] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 568.976358] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 569.059528] env[61936]: DEBUG nova.policy [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06199ff338204a119f8402413843e22d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fa23de03fc343fa94b8c58e2f928f34', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 569.254997] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.477641] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 569.631303] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c323e2a-334a-4599-924c-1182e2ddbebe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.639584] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3668a0-7c4a-4b50-9963-e6d7193a71ee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.673950] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac42fb8-3ac0-4540-a04c-273698fb652b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.682319] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a70f58-82c2-4e2f-92ae-89f23bbfee10 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.698364] env[61936]: DEBUG nova.compute.provider_tree [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.205200] env[61936]: DEBUG nova.scheduler.client.report [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 570.245887] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Successfully created port: dec513a9-aac4-4e9f-ba67-271187b72279 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 570.491263] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 570.524048] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 570.524301] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.524458] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 570.527281] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.527637] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 570.527877] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 570.528168] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 570.528556] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 570.528622] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 570.528822] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 570.529284] env[61936]: DEBUG nova.virt.hardware [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 570.530296] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7232e8-d45e-4d24-9c64-6505a3b32487 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.539399] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e0c548-1b74-417b-bb0e-92bef2941412 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.709647] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.710261] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 570.713371] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.615s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.715078] env[61936]: INFO nova.compute.claims [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 571.215152] env[61936]: DEBUG nova.compute.utils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.219811] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 571.219811] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 571.242250] env[61936]: ERROR nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. [ 571.242250] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 571.242250] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 571.242250] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 571.242250] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.242250] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 571.242250] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.242250] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 571.242250] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.242250] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 571.242250] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.242250] env[61936]: ERROR nova.compute.manager raise self.value [ 571.242250] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.242250] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 571.242250] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.242250] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 571.242762] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.242762] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 571.242762] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. [ 571.242762] env[61936]: ERROR nova.compute.manager [ 571.242762] env[61936]: Traceback (most recent call last): [ 571.242762] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 571.242762] env[61936]: listener.cb(fileno) [ 571.242762] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.242762] env[61936]: result = function(*args, **kwargs) [ 571.242762] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.242762] env[61936]: return func(*args, **kwargs) [ 571.242762] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 571.242762] env[61936]: raise e [ 571.242762] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 571.242762] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 571.242762] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.242762] env[61936]: created_port_ids = self._update_ports_for_instance( [ 571.242762] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.242762] env[61936]: with excutils.save_and_reraise_exception(): [ 571.242762] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.242762] env[61936]: self.force_reraise() [ 571.242762] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.242762] env[61936]: raise self.value [ 571.242762] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.242762] env[61936]: updated_port = self._update_port( [ 571.242762] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.242762] env[61936]: _ensure_no_port_binding_failure(port) [ 571.242762] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.242762] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 571.243596] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. [ 571.243596] env[61936]: Removing descriptor: 15 [ 571.245292] env[61936]: ERROR nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Traceback (most recent call last): [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] yield resources [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self.driver.spawn(context, instance, image_meta, [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] vm_ref = self.build_virtual_machine(instance, [ 571.245292] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] for vif in network_info: [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return self._sync_wrapper(fn, *args, **kwargs) [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self.wait() [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self[:] = self._gt.wait() [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return self._exit_event.wait() [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.245679] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] result = hub.switch() [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return self.greenlet.switch() [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] result = function(*args, **kwargs) [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return func(*args, **kwargs) [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] raise e [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] nwinfo = self.network_api.allocate_for_instance( [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] created_port_ids = self._update_ports_for_instance( [ 571.246407] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] with excutils.save_and_reraise_exception(): [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self.force_reraise() [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] raise self.value [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] updated_port = self._update_port( [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] _ensure_no_port_binding_failure(port) [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] raise exception.PortBindingFailed(port_id=port['id']) [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] nova.exception.PortBindingFailed: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. [ 571.246973] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] [ 571.247645] env[61936]: INFO nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Terminating instance [ 571.414682] env[61936]: DEBUG nova.policy [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '956066f01fcf4a08a1a35fe868ea5768', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '965f89c706fd4b6081a237e74b8ed1f9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.721212] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 571.755334] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Acquiring lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.755334] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Acquired lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.755334] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 571.876571] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f066021-e62d-46ec-8ac5-43185b2de85d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.885363] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e673ff-31c2-4dd0-baee-5e897ad3e781 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.922553] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e590a6-5797-4a97-a41f-5811536807e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.931213] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26aafd66-d9d0-41f6-a707-8db31f0ff261 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.945654] env[61936]: DEBUG nova.compute.provider_tree [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.191354] env[61936]: DEBUG nova.compute.manager [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Received event network-changed-37e2f9e8-3071-4377-95ab-e032be86f90e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 572.191354] env[61936]: DEBUG nova.compute.manager [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Refreshing instance network info cache due to event network-changed-37e2f9e8-3071-4377-95ab-e032be86f90e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 572.191354] env[61936]: DEBUG oslo_concurrency.lockutils [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] Acquiring lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.305144] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.449049] env[61936]: DEBUG nova.scheduler.client.report [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 572.515089] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Successfully created port: 3c8b50ba-50ab-4feb-bde4-dc4b83184616 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.662544] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.733099] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 572.764121] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 572.764121] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.764121] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 572.764285] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.764285] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 572.765167] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 572.765833] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 572.765833] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 572.765942] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 572.766171] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 572.766414] env[61936]: DEBUG nova.virt.hardware [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 572.767685] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efba5190-2592-4a35-a9ae-99434696385b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.776282] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3326527b-2525-44ea-8ccd-e3671e72de87 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.955714] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.955714] env[61936]: DEBUG nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 572.957780] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.794s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.960308] env[61936]: INFO nova.compute.claims [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.166987] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Releasing lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.167478] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 573.167661] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.167967] env[61936]: DEBUG oslo_concurrency.lockutils [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] Acquired lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.168150] env[61936]: DEBUG nova.network.neutron [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Refreshing network info cache for port 37e2f9e8-3071-4377-95ab-e032be86f90e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 573.169457] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b46ae11-b0c6-4ca6-b53e-f2e489d6ef52 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.182332] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf02ba70-dbc0-4c60-8595-cd71fc472dd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.210943] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bc67f44d-e6c8-49e0-ac2a-cf7449253580 could not be found. [ 573.211186] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 573.211565] env[61936]: INFO nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Took 0.04 seconds to destroy the instance on the hypervisor. [ 573.211820] env[61936]: DEBUG oslo.service.loopingcall [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.212047] env[61936]: DEBUG nova.compute.manager [-] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 573.212192] env[61936]: DEBUG nova.network.neutron [-] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.261830] env[61936]: DEBUG nova.network.neutron [-] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.466175] env[61936]: DEBUG nova.compute.utils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.474009] env[61936]: DEBUG nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Not allocating networking since 'none' was specified. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 573.529252] env[61936]: ERROR nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. [ 573.529252] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.529252] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.529252] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.529252] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.529252] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.529252] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.529252] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.529252] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.529252] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 573.529252] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.529252] env[61936]: ERROR nova.compute.manager raise self.value [ 573.529252] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.529252] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.529252] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.529252] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.529649] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.529649] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.529649] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. [ 573.529649] env[61936]: ERROR nova.compute.manager [ 573.531373] env[61936]: Traceback (most recent call last): [ 573.531373] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.531373] env[61936]: listener.cb(fileno) [ 573.531373] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.531373] env[61936]: result = function(*args, **kwargs) [ 573.531373] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.531373] env[61936]: return func(*args, **kwargs) [ 573.531373] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 573.531373] env[61936]: raise e [ 573.531373] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.531373] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 573.531373] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.531373] env[61936]: created_port_ids = self._update_ports_for_instance( [ 573.531373] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.531373] env[61936]: with excutils.save_and_reraise_exception(): [ 573.531373] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.531373] env[61936]: self.force_reraise() [ 573.531373] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.531373] env[61936]: raise self.value [ 573.531373] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.531373] env[61936]: updated_port = self._update_port( [ 573.531373] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.531373] env[61936]: _ensure_no_port_binding_failure(port) [ 573.531373] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.531373] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.531373] env[61936]: nova.exception.PortBindingFailed: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. [ 573.531373] env[61936]: Removing descriptor: 16 [ 573.532386] env[61936]: ERROR nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Traceback (most recent call last): [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] yield resources [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self.driver.spawn(context, instance, image_meta, [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] vm_ref = self.build_virtual_machine(instance, [ 573.532386] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] for vif in network_info: [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return self._sync_wrapper(fn, *args, **kwargs) [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self.wait() [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self[:] = self._gt.wait() [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return self._exit_event.wait() [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.532991] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] result = hub.switch() [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return self.greenlet.switch() [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] result = function(*args, **kwargs) [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return func(*args, **kwargs) [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] raise e [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] nwinfo = self.network_api.allocate_for_instance( [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] created_port_ids = self._update_ports_for_instance( [ 573.533479] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] with excutils.save_and_reraise_exception(): [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self.force_reraise() [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] raise self.value [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] updated_port = self._update_port( [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] _ensure_no_port_binding_failure(port) [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] raise exception.PortBindingFailed(port_id=port['id']) [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] nova.exception.PortBindingFailed: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. [ 573.534317] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] [ 573.534956] env[61936]: INFO nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Terminating instance [ 573.739499] env[61936]: DEBUG nova.network.neutron [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.764819] env[61936]: DEBUG nova.network.neutron [-] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.975420] env[61936]: DEBUG nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 574.038197] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.038402] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquired lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.038573] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.064290] env[61936]: DEBUG nova.network.neutron [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.183200] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17659ae1-4661-4f3b-8994-199433c75d5d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.193247] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bd27db-dc6d-4f0b-b0ba-8b685e1184ee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.249567] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68237515-5377-42b0-830b-3ac50a9bcf8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.259040] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910a4c53-06df-431e-8292-2a6f376c0efc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.277150] env[61936]: INFO nova.compute.manager [-] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Took 1.06 seconds to deallocate network for instance. [ 574.277150] env[61936]: DEBUG nova.compute.provider_tree [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.281302] env[61936]: DEBUG nova.compute.claims [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 574.281302] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.569195] env[61936]: DEBUG oslo_concurrency.lockutils [req-6136ff0c-4250-4af0-96e3-aff6f2812306 req-fd039fbc-8279-4ab7-a3c0-6e10ae2c814d service nova] Releasing lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.589102] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.753854] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.754297] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.775601] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.781342] env[61936]: DEBUG nova.scheduler.client.report [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 574.994360] env[61936]: DEBUG nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 575.025080] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 575.025639] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.025639] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 575.025740] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.025811] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 575.025946] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 575.026356] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 575.026974] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 575.027223] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 575.027388] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 575.027568] env[61936]: DEBUG nova.virt.hardware [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 575.029782] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6a4318-e04f-4131-9d88-f71027bab02f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.039161] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba0f799-80d5-402b-a208-f73641e23563 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.057256] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 575.068654] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 575.068879] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53a76446-e12c-4e20-82e3-7d74ed97f55e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.081847] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Created folder: OpenStack in parent group-v4. [ 575.082049] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Creating folder: Project (265ab171bc7046e4b244897f4be61a28). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 575.082289] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0da5756-7d01-4052-b307-0c0adc42f9ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.091717] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Created folder: Project (265ab171bc7046e4b244897f4be61a28) in parent group-v269874. [ 575.091909] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Creating folder: Instances. Parent ref: group-v269875. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 575.092157] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a27c82d0-83a4-4d10-a10a-c7e9c428ca6d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.104445] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Created folder: Instances in parent group-v269875. [ 575.104775] env[61936]: DEBUG oslo.service.loopingcall [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.105033] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 575.105197] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f6e0d05-59a7-44e9-a47b-44834460c9d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.125440] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 575.125440] env[61936]: value = "task-1252693" [ 575.125440] env[61936]: _type = "Task" [ 575.125440] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.141500] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252693, 'name': CreateVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.261259] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 575.282026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Releasing lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.282026] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 575.282026] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.282026] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-65a14d82-716d-4817-8252-403f766365ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.291899] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72d4d7d-67f9-4d63-8c8f-49396a102600 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.304098] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.304932] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 575.307118] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.755s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.312019] env[61936]: INFO nova.compute.claims [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.329636] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c4b372a3-f76a-411d-a336-338f5c6d502c could not be found. [ 575.329636] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 575.329636] env[61936]: INFO nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 575.329636] env[61936]: DEBUG oslo.service.loopingcall [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.334018] env[61936]: DEBUG nova.compute.manager [-] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 575.334018] env[61936]: DEBUG nova.network.neutron [-] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.432871] env[61936]: DEBUG nova.network.neutron [-] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.464561] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.464885] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.465185] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 575.465185] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Rebuilding the list of instances to heal {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10330}} [ 575.643402] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252693, 'name': CreateVM_Task, 'duration_secs': 0.307123} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.643402] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 575.644926] env[61936]: DEBUG oslo_vmware.service [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f059a37-3187-43b2-aeec-e897a2f97983 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.653153] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.653325] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.654568] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 575.656281] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0aea48e4-8a0a-433a-a6a4-6c986edfe2c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.663340] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 575.663340] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cb5ab4-ace6-7ccd-28a0-56ca92579e88" [ 575.663340] env[61936]: _type = "Task" [ 575.663340] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.673581] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cb5ab4-ace6-7ccd-28a0-56ca92579e88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.686730] env[61936]: ERROR nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. [ 575.686730] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.686730] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.686730] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.686730] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.686730] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.686730] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.686730] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.686730] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.686730] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 575.686730] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.686730] env[61936]: ERROR nova.compute.manager raise self.value [ 575.686730] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.686730] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.686730] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.686730] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.687308] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.687308] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.687308] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. [ 575.687308] env[61936]: ERROR nova.compute.manager [ 575.687308] env[61936]: Traceback (most recent call last): [ 575.687308] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.687308] env[61936]: listener.cb(fileno) [ 575.687308] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.687308] env[61936]: result = function(*args, **kwargs) [ 575.687308] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.687308] env[61936]: return func(*args, **kwargs) [ 575.687308] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 575.687308] env[61936]: raise e [ 575.687308] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.687308] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 575.687308] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.687308] env[61936]: created_port_ids = self._update_ports_for_instance( [ 575.687308] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.687308] env[61936]: with excutils.save_and_reraise_exception(): [ 575.687308] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.687308] env[61936]: self.force_reraise() [ 575.687308] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.687308] env[61936]: raise self.value [ 575.687308] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.687308] env[61936]: updated_port = self._update_port( [ 575.687308] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.687308] env[61936]: _ensure_no_port_binding_failure(port) [ 575.687308] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.687308] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.688870] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. [ 575.688870] env[61936]: Removing descriptor: 17 [ 575.688870] env[61936]: ERROR nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Traceback (most recent call last): [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] yield resources [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self.driver.spawn(context, instance, image_meta, [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.688870] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] vm_ref = self.build_virtual_machine(instance, [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] for vif in network_info: [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return self._sync_wrapper(fn, *args, **kwargs) [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self.wait() [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self[:] = self._gt.wait() [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return self._exit_event.wait() [ 575.689183] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] result = hub.switch() [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return self.greenlet.switch() [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] result = function(*args, **kwargs) [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return func(*args, **kwargs) [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] raise e [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] nwinfo = self.network_api.allocate_for_instance( [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.689517] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] created_port_ids = self._update_ports_for_instance( [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] with excutils.save_and_reraise_exception(): [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self.force_reraise() [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] raise self.value [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] updated_port = self._update_port( [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] _ensure_no_port_binding_failure(port) [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.689811] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] raise exception.PortBindingFailed(port_id=port['id']) [ 575.690756] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] nova.exception.PortBindingFailed: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. [ 575.690756] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] [ 575.690756] env[61936]: INFO nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Terminating instance [ 575.788844] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.816915] env[61936]: DEBUG nova.compute.utils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.822741] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 575.822920] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 575.855127] env[61936]: DEBUG nova.compute.manager [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Received event network-vif-deleted-37e2f9e8-3071-4377-95ab-e032be86f90e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 575.855263] env[61936]: DEBUG nova.compute.manager [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Received event network-changed-dec513a9-aac4-4e9f-ba67-271187b72279 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 575.855340] env[61936]: DEBUG nova.compute.manager [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Refreshing instance network info cache due to event network-changed-dec513a9-aac4-4e9f-ba67-271187b72279. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 575.855534] env[61936]: DEBUG oslo_concurrency.lockutils [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] Acquiring lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.855699] env[61936]: DEBUG oslo_concurrency.lockutils [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] Acquired lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.857815] env[61936]: DEBUG nova.network.neutron [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Refreshing network info cache for port dec513a9-aac4-4e9f-ba67-271187b72279 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 575.936279] env[61936]: DEBUG nova.network.neutron [-] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.971720] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 575.971891] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 575.972027] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 575.972203] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 575.972261] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 575.972380] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 575.972686] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Didn't find any instances for network info cache update. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10412}} [ 575.972812] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.973250] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.973250] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.973443] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.973772] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.974028] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 575.974176] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 575.974433] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 576.055770] env[61936]: DEBUG nova.policy [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1516c349d95c4ebd8ed1a12f810d2104', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '242f53cc5d1a4f67bc439d7c7a198663', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.175619] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.175872] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 576.176203] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.176274] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.176651] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 576.176897] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34d60be3-86be-47b1-80de-37b932952b5e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.193871] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Acquiring lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.194109] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Acquired lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.194237] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.197481] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 576.197481] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 576.198597] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8bddd1-37fb-4790-8b3d-d4f6aac33a6a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.205703] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d1b4f82-065a-40bd-93c5-d76fde97f6a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.211833] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 576.211833] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5200d64b-b4e3-2d96-d081-8fcb7d45f4bd" [ 576.211833] env[61936]: _type = "Task" [ 576.211833] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.222409] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5200d64b-b4e3-2d96-d081-8fcb7d45f4bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.324166] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 576.409554] env[61936]: DEBUG nova.network.neutron [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.439851] env[61936]: INFO nova.compute.manager [-] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Took 1.11 seconds to deallocate network for instance. [ 576.441950] env[61936]: DEBUG nova.compute.claims [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 576.442201] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.481386] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.547711] env[61936]: DEBUG nova.compute.manager [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Received event network-changed-3c8b50ba-50ab-4feb-bde4-dc4b83184616 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 576.547908] env[61936]: DEBUG nova.compute.manager [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Refreshing instance network info cache due to event network-changed-3c8b50ba-50ab-4feb-bde4-dc4b83184616. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 576.548104] env[61936]: DEBUG oslo_concurrency.lockutils [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] Acquiring lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.580999] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63b3056-1ef8-4d50-bfbf-009728b43e44 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.586279] env[61936]: DEBUG nova.network.neutron [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.592930] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909e79e6-9914-4e5c-91e8-e9ff39ba544b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.637735] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddbc3cb-b23d-4127-b5f6-24810c6194ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.646427] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95a8e6b-4074-4401-a035-9ae7e0e22c2c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.660992] env[61936]: DEBUG nova.compute.provider_tree [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.724341] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Preparing fetch location {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 576.724607] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Creating directory with path [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 576.724836] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20ba5d08-a13b-45a3-9d73-777479c5aaa2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.736353] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.745238] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Created directory with path [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 576.745448] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Fetch image to [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 576.745613] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Downloading image file data 43acc3d3-5e18-42a0-9168-cb6831c6bbfb to [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk on the data store datastore1 {{(pid=61936) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 576.746439] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa52457-31c2-449c-a6c8-f2a2be198f4c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.753866] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd286ed4-effe-4462-8052-82e890364fd8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.765498] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c525cf7-9608-4971-a5e8-641f9061cb8e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.806188] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e05ef4-41bc-42e7-876b-37846f3051ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.812525] env[61936]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-51cea4c3-d806-450b-b820-a23c3ff2914a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.842112] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Downloading image file data 43acc3d3-5e18-42a0-9168-cb6831c6bbfb to the data store datastore1 {{(pid=61936) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 576.924217] env[61936]: DEBUG oslo_vmware.rw_handles [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61936) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 576.926045] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.094179] env[61936]: DEBUG oslo_concurrency.lockutils [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] Releasing lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.094179] env[61936]: DEBUG nova.compute.manager [req-902127bf-135d-4757-950b-049b77e01a27 req-52749093-9888-4513-a6b0-b0cb2e38d767 service nova] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Received event network-vif-deleted-dec513a9-aac4-4e9f-ba67-271187b72279 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 577.120303] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "1944e478-e7ff-4c20-bdfe-95016613c704" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.120544] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "1944e478-e7ff-4c20-bdfe-95016613c704" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.166993] env[61936]: DEBUG nova.scheduler.client.report [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 577.195344] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Successfully created port: 383074b9-b966-443d-a4fc-519c9725d03c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 577.338787] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 577.395266] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 577.395692] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.395692] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 577.399564] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.399564] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 577.399564] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 577.399564] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 577.399564] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 577.399758] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 577.399758] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 577.399758] env[61936]: DEBUG nova.virt.hardware [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 577.399758] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66cb872-7dfd-4a1e-b701-8b19583dddba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.412019] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb563e1-56fa-4c00-8ec6-454a1e15851d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.435474] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Releasing lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.436286] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 577.436404] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 577.439091] env[61936]: DEBUG oslo_concurrency.lockutils [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] Acquired lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.439277] env[61936]: DEBUG nova.network.neutron [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Refreshing network info cache for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 577.440952] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fa7951a-6c5b-496a-b96e-5385ef8edeb1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.453402] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44823c31-b086-4113-85a7-4dcabedb3ed6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.486996] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9 could not be found. [ 577.486996] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 577.488916] env[61936]: INFO nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 577.489061] env[61936]: DEBUG oslo.service.loopingcall [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.491217] env[61936]: DEBUG nova.compute.manager [-] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 577.491310] env[61936]: DEBUG nova.network.neutron [-] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.557516] env[61936]: DEBUG nova.network.neutron [-] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.631203] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 577.680975] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.680975] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 577.683362] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.522s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.684844] env[61936]: INFO nova.compute.claims [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.753052] env[61936]: DEBUG oslo_vmware.rw_handles [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Completed reading data from the image iterator. {{(pid=61936) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 577.753302] env[61936]: DEBUG oslo_vmware.rw_handles [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 577.757934] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Acquiring lock "0257bab6-c040-4f1f-8218-c5bc353777bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.758538] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Lock "0257bab6-c040-4f1f-8218-c5bc353777bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.847833] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Downloaded image file data 43acc3d3-5e18-42a0-9168-cb6831c6bbfb to vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk on the data store datastore1 {{(pid=61936) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 577.849795] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Caching image {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 577.850084] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Copying Virtual Disk [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk to [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 577.850408] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88d2f82a-644b-47df-8aed-4adc317bdd87 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.858955] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 577.858955] env[61936]: value = "task-1252694" [ 577.858955] env[61936]: _type = "Task" [ 577.858955] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.867660] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252694, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.020228] env[61936]: DEBUG nova.network.neutron [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.062310] env[61936]: DEBUG nova.network.neutron [-] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.163285] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.189471] env[61936]: DEBUG nova.compute.utils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.195491] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 578.195491] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 578.308096] env[61936]: DEBUG nova.network.neutron [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.377571] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252694, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.517816] env[61936]: DEBUG nova.policy [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5ff6a17c53844aea1e6d518d6b2ac9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d12e11633bd4d3ebdaefd62587ca4ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.569051] env[61936]: INFO nova.compute.manager [-] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Took 1.08 seconds to deallocate network for instance. [ 578.571530] env[61936]: DEBUG nova.compute.claims [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.571706] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.697427] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 578.817256] env[61936]: DEBUG oslo_concurrency.lockutils [req-95dd6a38-1aaa-49c4-88f7-340653e0a1ca req-c0f634ad-2311-4df1-a65a-6181b8069cbc service nova] Releasing lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.873854] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252694, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.713231} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.876618] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Copied Virtual Disk [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk to [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 578.876831] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Deleting the datastore file [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 578.877517] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2099545d-97f7-4926-8c84-56a5cdd76239 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.886049] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 578.886049] env[61936]: value = "task-1252695" [ 578.886049] env[61936]: _type = "Task" [ 578.886049] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.892510] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252695, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.940439] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057add92-bf08-4df2-aa07-5b3d99696046 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.947658] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a951977-55d3-4bd7-a4a9-2043e00bda7a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.983497] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9736d789-8147-436e-8d04-4a0349fe92af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.991689] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8b9699-56c2-4ace-832d-0a1ad5592354 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.006174] env[61936]: DEBUG nova.compute.provider_tree [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.062717] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.063495] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.397512] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252695, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027711} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.397512] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 579.397512] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Moving file from [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4/43acc3d3-5e18-42a0-9168-cb6831c6bbfb to [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb. {{(pid=61936) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 579.397660] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-09b966da-d2a9-4dd8-aa18-94ff65a03a4b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.405304] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 579.405304] env[61936]: value = "task-1252696" [ 579.405304] env[61936]: _type = "Task" [ 579.405304] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.413877] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252696, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.512843] env[61936]: DEBUG nova.scheduler.client.report [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 579.556135] env[61936]: DEBUG nova.compute.manager [req-36c126ae-c8a1-4c0d-9766-a0c86e76241d req-85e83ea8-dcb5-4f7b-8076-3c290c22acb4 service nova] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Received event network-vif-deleted-3c8b50ba-50ab-4feb-bde4-dc4b83184616 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 579.714303] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 579.744816] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 579.748247] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.748453] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 579.748640] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.748777] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 579.748915] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 579.749144] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 579.749293] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 579.749448] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 579.749595] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 579.749758] env[61936]: DEBUG nova.virt.hardware [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 579.751096] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c2afdb-5274-4877-a6cc-e66bb6924a7b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.759990] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5880006-4d37-44af-b36a-e5d59b2b94f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.915731] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252696, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025508} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.916418] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] File moved {{(pid=61936) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 579.916747] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Cleaning up location [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4 {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 579.917189] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Deleting the datastore file [datastore1] vmware_temp/cfe8b92e-aced-46d6-91ad-9e3fa930e2d4 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 579.917535] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-352d5e29-017b-4be3-868e-e3303de15c56 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.924932] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 579.924932] env[61936]: value = "task-1252697" [ 579.924932] env[61936]: _type = "Task" [ 579.924932] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.934015] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252697, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.017182] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.020598] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 580.025240] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.769s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.025240] env[61936]: INFO nova.compute.claims [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.306870] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Successfully created port: 977a9bb1-444c-48a0-a662-fb19f32b5e72 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.438101] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252697, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026052} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.438860] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 580.441022] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f73eccf-eb5d-43c8-9fa5-67c1ab4a9d77 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.446214] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 580.446214] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d78ea8-a686-f146-4c92-66d179ca290e" [ 580.446214] env[61936]: _type = "Task" [ 580.446214] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.455426] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d78ea8-a686-f146-4c92-66d179ca290e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.532060] env[61936]: DEBUG nova.compute.utils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.533613] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 580.533818] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 580.811457] env[61936]: DEBUG nova.policy [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7fdd8324d444ad4b80b6eb6b05f9984', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f5ba82acd64284966710a7f991d509', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.959280] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d78ea8-a686-f146-4c92-66d179ca290e, 'name': SearchDatastore_Task, 'duration_secs': 0.009706} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.959520] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.959921] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 7dadef86-7685-40e2-9e68-debc4b4a8a95/7dadef86-7685-40e2-9e68-debc4b4a8a95.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 580.960061] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81ee8e90-c8e2-48ec-ac04-ba0e71bd6273 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.971116] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 580.971116] env[61936]: value = "task-1252698" [ 580.971116] env[61936]: _type = "Task" [ 580.971116] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.981492] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.039333] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 581.194184] env[61936]: ERROR nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. [ 581.194184] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.194184] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 581.194184] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.194184] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.194184] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.194184] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.194184] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.194184] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.194184] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 581.194184] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.194184] env[61936]: ERROR nova.compute.manager raise self.value [ 581.194184] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.194184] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.194184] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.194184] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.194973] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.194973] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.194973] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. [ 581.194973] env[61936]: ERROR nova.compute.manager [ 581.194973] env[61936]: Traceback (most recent call last): [ 581.194973] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.194973] env[61936]: listener.cb(fileno) [ 581.194973] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.194973] env[61936]: result = function(*args, **kwargs) [ 581.194973] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.194973] env[61936]: return func(*args, **kwargs) [ 581.194973] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 581.194973] env[61936]: raise e [ 581.194973] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 581.194973] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 581.194973] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.194973] env[61936]: created_port_ids = self._update_ports_for_instance( [ 581.194973] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.194973] env[61936]: with excutils.save_and_reraise_exception(): [ 581.194973] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.194973] env[61936]: self.force_reraise() [ 581.194973] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.194973] env[61936]: raise self.value [ 581.194973] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.194973] env[61936]: updated_port = self._update_port( [ 581.194973] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.194973] env[61936]: _ensure_no_port_binding_failure(port) [ 581.194973] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.194973] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.196456] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. [ 581.196456] env[61936]: Removing descriptor: 17 [ 581.196456] env[61936]: ERROR nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Traceback (most recent call last): [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] yield resources [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self.driver.spawn(context, instance, image_meta, [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.196456] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] vm_ref = self.build_virtual_machine(instance, [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] for vif in network_info: [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return self._sync_wrapper(fn, *args, **kwargs) [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self.wait() [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self[:] = self._gt.wait() [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return self._exit_event.wait() [ 581.196845] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] result = hub.switch() [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return self.greenlet.switch() [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] result = function(*args, **kwargs) [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return func(*args, **kwargs) [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] raise e [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] nwinfo = self.network_api.allocate_for_instance( [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.198216] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] created_port_ids = self._update_ports_for_instance( [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] with excutils.save_and_reraise_exception(): [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self.force_reraise() [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] raise self.value [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] updated_port = self._update_port( [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] _ensure_no_port_binding_failure(port) [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.198723] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] raise exception.PortBindingFailed(port_id=port['id']) [ 581.199183] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] nova.exception.PortBindingFailed: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. [ 581.199183] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] [ 581.199183] env[61936]: INFO nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Terminating instance [ 581.288504] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c0003b-c63e-4db2-812c-d90573222a48 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.298335] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238efa75-720d-4cd2-a74e-4a388653312e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.340500] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a72b8a-de3d-49ae-b253-1bf6008d66c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.351826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Acquiring lock "39a94d7e-77bb-4083-81c3-4dce3bf55fdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.352123] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Lock "39a94d7e-77bb-4083-81c3-4dce3bf55fdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.361088] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c0e31e-6ed1-4ae0-87ad-9a0a09eb7e70 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.378960] env[61936]: DEBUG nova.compute.provider_tree [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.483726] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461409} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.483726] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 7dadef86-7685-40e2-9e68-debc4b4a8a95/7dadef86-7685-40e2-9e68-debc4b4a8a95.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 581.483726] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 581.483726] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-140153ae-c0e8-461e-891f-cde17058f1ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.489553] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 581.489553] env[61936]: value = "task-1252699" [ 581.489553] env[61936]: _type = "Task" [ 581.489553] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.499915] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252699, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.701909] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Acquiring lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.702019] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Acquired lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.702288] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.883575] env[61936]: DEBUG nova.scheduler.client.report [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 582.001129] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252699, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186409} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.001944] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 582.002595] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4782ecfd-0f67-403a-8d6b-5c28f7a99b1a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.027904] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 7dadef86-7685-40e2-9e68-debc4b4a8a95/7dadef86-7685-40e2-9e68-debc4b4a8a95.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 582.028432] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0731746-54ba-4a9f-9b96-6e994a09d837 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.056624] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 582.067094] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 582.067094] env[61936]: value = "task-1252700" [ 582.067094] env[61936]: _type = "Task" [ 582.067094] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.078534] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252700, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.108454] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 582.108729] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.108882] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 582.110309] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.110309] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 582.110309] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 582.110309] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 582.110733] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 582.110733] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 582.110917] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 582.111310] env[61936]: DEBUG nova.virt.hardware [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 582.113049] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afe7275-8377-4322-a9bb-fb0bef5fc6e5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.123989] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24573fc-ed9d-4fd5-98bb-9be72582377b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.253161] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.392296] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.393660] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 582.399943] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.116s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.469477] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Successfully created port: 9fcab735-be3b-48c2-a158-50e091ace076 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.539667] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.582831] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252700, 'name': ReconfigVM_Task, 'duration_secs': 0.312108} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.583116] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 7dadef86-7685-40e2-9e68-debc4b4a8a95/7dadef86-7685-40e2-9e68-debc4b4a8a95.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 582.583809] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c34bc713-1c20-4e8b-9bfa-e3fcd9f7f737 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.590611] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 582.590611] env[61936]: value = "task-1252701" [ 582.590611] env[61936]: _type = "Task" [ 582.590611] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.599742] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252701, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.823403] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "aad2d5db-9e34-40d3-816c-902eecd0069e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.823746] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "aad2d5db-9e34-40d3-816c-902eecd0069e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.905077] env[61936]: DEBUG nova.compute.utils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.910041] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 582.911043] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 583.044660] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Releasing lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.045204] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 583.045416] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.045724] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31567fbd-ef47-4722-baf0-9560e464e337 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.059134] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67d163f-98be-4dbb-a66e-1dc9b9f7eb23 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.086425] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2 could not be found. [ 583.086673] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 583.086878] env[61936]: INFO nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 583.087235] env[61936]: DEBUG oslo.service.loopingcall [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.091717] env[61936]: DEBUG nova.compute.manager [-] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 583.091717] env[61936]: DEBUG nova.network.neutron [-] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.097170] env[61936]: DEBUG nova.policy [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '335f4520b78f404ca67166af9f676ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca412a547c0845bb95ecf0a9f14effb1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.104561] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252701, 'name': Rename_Task, 'duration_secs': 0.140648} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.104977] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 583.105284] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c12813af-7bb7-4729-b126-2e5c03f62cc0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.113032] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 583.113032] env[61936]: value = "task-1252702" [ 583.113032] env[61936]: _type = "Task" [ 583.113032] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.121919] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252702, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.130047] env[61936]: DEBUG nova.network.neutron [-] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.172688] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4b0ebf-68d6-49fc-8c41-ffd757b401ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.181634] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05ef65c-42ec-469c-9215-5a45d6cf638c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.216318] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5598aac3-65a7-4d17-bff3-5ec6b94c17e8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.220035] env[61936]: DEBUG nova.compute.manager [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Received event network-changed-383074b9-b966-443d-a4fc-519c9725d03c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 583.220035] env[61936]: DEBUG nova.compute.manager [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Refreshing instance network info cache due to event network-changed-383074b9-b966-443d-a4fc-519c9725d03c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 583.220131] env[61936]: DEBUG oslo_concurrency.lockutils [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] Acquiring lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.220196] env[61936]: DEBUG oslo_concurrency.lockutils [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] Acquired lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.220333] env[61936]: DEBUG nova.network.neutron [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Refreshing network info cache for port 383074b9-b966-443d-a4fc-519c9725d03c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.228543] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493a0e35-d210-4fd7-a59a-e18198bf85db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.243566] env[61936]: DEBUG nova.compute.provider_tree [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.413393] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 583.625122] env[61936]: DEBUG oslo_vmware.api [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252702, 'name': PowerOnVM_Task, 'duration_secs': 0.505704} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.625403] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 583.625544] env[61936]: INFO nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Took 8.63 seconds to spawn the instance on the hypervisor. [ 583.625791] env[61936]: DEBUG nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 583.626676] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48d9662-12d7-4e5c-a475-e86020244e60 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.631785] env[61936]: DEBUG nova.network.neutron [-] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.746422] env[61936]: DEBUG nova.scheduler.client.report [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 583.761888] env[61936]: DEBUG nova.network.neutron [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.948661] env[61936]: DEBUG nova.network.neutron [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.097784] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Acquiring lock "de5277a7-05e2-49eb-b577-87f900eeef3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.100423] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Lock "de5277a7-05e2-49eb-b577-87f900eeef3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.119255] env[61936]: ERROR nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. [ 584.119255] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.119255] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.119255] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.119255] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.119255] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.119255] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.119255] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.119255] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.119255] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 584.119255] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.119255] env[61936]: ERROR nova.compute.manager raise self.value [ 584.119255] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.119255] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.119255] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.119255] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.119735] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.119735] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.119735] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. [ 584.119735] env[61936]: ERROR nova.compute.manager [ 584.119735] env[61936]: Traceback (most recent call last): [ 584.119735] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.119735] env[61936]: listener.cb(fileno) [ 584.119735] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.119735] env[61936]: result = function(*args, **kwargs) [ 584.119735] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.119735] env[61936]: return func(*args, **kwargs) [ 584.119735] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 584.119735] env[61936]: raise e [ 584.119735] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.119735] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 584.119735] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.119735] env[61936]: created_port_ids = self._update_ports_for_instance( [ 584.119735] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.119735] env[61936]: with excutils.save_and_reraise_exception(): [ 584.119735] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.119735] env[61936]: self.force_reraise() [ 584.119735] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.119735] env[61936]: raise self.value [ 584.119735] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.119735] env[61936]: updated_port = self._update_port( [ 584.119735] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.119735] env[61936]: _ensure_no_port_binding_failure(port) [ 584.119735] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.119735] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.121329] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. [ 584.121329] env[61936]: Removing descriptor: 15 [ 584.121329] env[61936]: ERROR nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Traceback (most recent call last): [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] yield resources [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self.driver.spawn(context, instance, image_meta, [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.121329] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] vm_ref = self.build_virtual_machine(instance, [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] for vif in network_info: [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return self._sync_wrapper(fn, *args, **kwargs) [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self.wait() [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self[:] = self._gt.wait() [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return self._exit_event.wait() [ 584.121807] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] result = hub.switch() [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return self.greenlet.switch() [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] result = function(*args, **kwargs) [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return func(*args, **kwargs) [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] raise e [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] nwinfo = self.network_api.allocate_for_instance( [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.122181] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] created_port_ids = self._update_ports_for_instance( [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] with excutils.save_and_reraise_exception(): [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self.force_reraise() [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] raise self.value [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] updated_port = self._update_port( [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] _ensure_no_port_binding_failure(port) [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.122633] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] raise exception.PortBindingFailed(port_id=port['id']) [ 584.122968] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] nova.exception.PortBindingFailed: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. [ 584.122968] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] [ 584.122968] env[61936]: INFO nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Terminating instance [ 584.138777] env[61936]: INFO nova.compute.manager [-] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Took 1.05 seconds to deallocate network for instance. [ 584.152593] env[61936]: INFO nova.compute.manager [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Took 17.09 seconds to build instance. [ 584.165691] env[61936]: DEBUG nova.compute.claims [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 584.165819] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.247776] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Successfully created port: fa16c095-21c9-4908-b9b7-2b5dfa7aaad5 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.257504] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.857s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.257504] env[61936]: ERROR nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. [ 584.257504] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Traceback (most recent call last): [ 584.257504] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 584.257504] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self.driver.spawn(context, instance, image_meta, [ 584.257504] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 584.257504] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.257504] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.257504] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] vm_ref = self.build_virtual_machine(instance, [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] for vif in network_info: [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return self._sync_wrapper(fn, *args, **kwargs) [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self.wait() [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self[:] = self._gt.wait() [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return self._exit_event.wait() [ 584.257754] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] result = hub.switch() [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return self.greenlet.switch() [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] result = function(*args, **kwargs) [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] return func(*args, **kwargs) [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] raise e [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] nwinfo = self.network_api.allocate_for_instance( [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.258067] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] created_port_ids = self._update_ports_for_instance( [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] with excutils.save_and_reraise_exception(): [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] self.force_reraise() [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] raise self.value [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] updated_port = self._update_port( [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] _ensure_no_port_binding_failure(port) [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.258422] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] raise exception.PortBindingFailed(port_id=port['id']) [ 584.258804] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] nova.exception.PortBindingFailed: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. [ 584.258804] env[61936]: ERROR nova.compute.manager [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] [ 584.258804] env[61936]: DEBUG nova.compute.utils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.258804] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.468s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.261028] env[61936]: INFO nova.compute.claims [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.272332] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Build of instance bc67f44d-e6c8-49e0-ac2a-cf7449253580 was re-scheduled: Binding failed for port 37e2f9e8-3071-4377-95ab-e032be86f90e, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 584.272332] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 584.272512] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Acquiring lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.277038] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Acquired lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.277038] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.424697] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 584.455205] env[61936]: DEBUG oslo_concurrency.lockutils [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] Releasing lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.460190] env[61936]: DEBUG nova.compute.manager [req-b77668a0-7786-4b84-8afc-9abfeaf100f7 req-dac4b840-48e0-4a27-a817-adea51caae3b service nova] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Received event network-vif-deleted-383074b9-b966-443d-a4fc-519c9725d03c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 584.467113] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 584.469111] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.469111] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 584.469111] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.469111] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 584.469111] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 584.469304] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 584.469304] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 584.469304] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 584.469304] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 584.469304] env[61936]: DEBUG nova.virt.hardware [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 584.469762] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87c31c8-ef7b-4640-8ee7-388c0ef6c355 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.480468] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56ed778-5d07-43bd-91cb-910ce6dc80b3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.626158] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Acquiring lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.626158] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Acquired lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.626333] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.659495] env[61936]: DEBUG oslo_concurrency.lockutils [None req-08262331-3da8-4611-92a9-081b276da008 tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "7dadef86-7685-40e2-9e68-debc4b4a8a95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.599s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.837327] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Acquiring lock "c9302959-c96c-47cb-980d-13493aa185f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.837609] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Lock "c9302959-c96c-47cb-980d-13493aa185f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.873999] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.954108] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Acquiring lock "69563e1e-301a-4a7b-8fe4-7f1846478f56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.954952] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Lock "69563e1e-301a-4a7b-8fe4-7f1846478f56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.162942] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 585.183023] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.262077] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.463995] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.645786] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b83c99-bd11-4517-bcf6-99aec0d6da51 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.660021] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e91ef1d-d1bd-4505-b1b3-525ec0a4d71c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.695850] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f80e3e8-edb2-4453-ba51-a9b01bd53a05 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.708596] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56831ce-8783-46e8-8ac3-3504836c2b05 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.726330] env[61936]: DEBUG nova.compute.provider_tree [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.729690] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.733041] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Acquiring lock "9ac0103e-8424-4518-8acb-343267deb3ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.733269] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Lock "9ac0103e-8424-4518-8acb-343267deb3ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.743524] env[61936]: ERROR nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. [ 585.743524] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.743524] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.743524] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.743524] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.743524] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.743524] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.743524] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.743524] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.743524] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 585.743524] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.743524] env[61936]: ERROR nova.compute.manager raise self.value [ 585.743524] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.743524] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.743524] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.743524] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.743984] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.743984] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.743984] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. [ 585.743984] env[61936]: ERROR nova.compute.manager [ 585.743984] env[61936]: Traceback (most recent call last): [ 585.743984] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.743984] env[61936]: listener.cb(fileno) [ 585.743984] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.743984] env[61936]: result = function(*args, **kwargs) [ 585.743984] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.743984] env[61936]: return func(*args, **kwargs) [ 585.743984] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.743984] env[61936]: raise e [ 585.743984] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.743984] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 585.743984] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.743984] env[61936]: created_port_ids = self._update_ports_for_instance( [ 585.743984] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.743984] env[61936]: with excutils.save_and_reraise_exception(): [ 585.743984] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.743984] env[61936]: self.force_reraise() [ 585.743984] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.743984] env[61936]: raise self.value [ 585.743984] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.743984] env[61936]: updated_port = self._update_port( [ 585.743984] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.743984] env[61936]: _ensure_no_port_binding_failure(port) [ 585.743984] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.743984] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.744776] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. [ 585.744776] env[61936]: Removing descriptor: 16 [ 585.744776] env[61936]: ERROR nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Traceback (most recent call last): [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] yield resources [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self.driver.spawn(context, instance, image_meta, [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.744776] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] vm_ref = self.build_virtual_machine(instance, [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] for vif in network_info: [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return self._sync_wrapper(fn, *args, **kwargs) [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self.wait() [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self[:] = self._gt.wait() [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return self._exit_event.wait() [ 585.745231] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] result = hub.switch() [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return self.greenlet.switch() [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] result = function(*args, **kwargs) [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return func(*args, **kwargs) [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] raise e [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] nwinfo = self.network_api.allocate_for_instance( [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.745618] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] created_port_ids = self._update_ports_for_instance( [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] with excutils.save_and_reraise_exception(): [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self.force_reraise() [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] raise self.value [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] updated_port = self._update_port( [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] _ensure_no_port_binding_failure(port) [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.746007] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] raise exception.PortBindingFailed(port_id=port['id']) [ 585.746411] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] nova.exception.PortBindingFailed: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. [ 585.746411] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] [ 585.746411] env[61936]: INFO nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Terminating instance [ 585.763027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Releasing lock "refresh_cache-bc67f44d-e6c8-49e0-ac2a-cf7449253580" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.763252] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 585.763546] env[61936]: DEBUG nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 585.763616] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.805942] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.967249] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Releasing lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.967680] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 585.967871] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 585.968199] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e4edbef-0dac-42f3-98b0-5b8c20f81b77 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.978272] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599065c4-d20f-4c35-a0c5-b460d234f1bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.000901] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5aa38a34-2f1c-4546-9688-980d65dfed27 could not be found. [ 586.001102] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.001344] env[61936]: INFO nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Took 0.03 seconds to destroy the instance on the hypervisor. [ 586.001784] env[61936]: DEBUG oslo.service.loopingcall [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.001784] env[61936]: DEBUG nova.compute.manager [-] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.001919] env[61936]: DEBUG nova.network.neutron [-] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.044235] env[61936]: DEBUG nova.network.neutron [-] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.178317] env[61936]: DEBUG nova.compute.manager [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Received event network-changed-977a9bb1-444c-48a0-a662-fb19f32b5e72 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 586.178781] env[61936]: DEBUG nova.compute.manager [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Refreshing instance network info cache due to event network-changed-977a9bb1-444c-48a0-a662-fb19f32b5e72. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 586.178781] env[61936]: DEBUG oslo_concurrency.lockutils [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] Acquiring lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.178847] env[61936]: DEBUG oslo_concurrency.lockutils [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] Acquired lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.178980] env[61936]: DEBUG nova.network.neutron [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Refreshing network info cache for port 977a9bb1-444c-48a0-a662-fb19f32b5e72 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 586.236605] env[61936]: DEBUG nova.scheduler.client.report [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 586.253832] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Acquiring lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.253832] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Acquired lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.253832] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.311465] env[61936]: DEBUG nova.network.neutron [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.546801] env[61936]: DEBUG nova.network.neutron [-] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.724098] env[61936]: DEBUG nova.network.neutron [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.742520] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.743692] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 586.751455] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.309s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.794806] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.815590] env[61936]: INFO nova.compute.manager [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] [instance: bc67f44d-e6c8-49e0-ac2a-cf7449253580] Took 1.05 seconds to deallocate network for instance. [ 586.846424] env[61936]: DEBUG nova.network.neutron [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.890692] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.052508] env[61936]: INFO nova.compute.manager [-] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Took 1.05 seconds to deallocate network for instance. [ 587.054207] env[61936]: DEBUG nova.compute.claims [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.054712] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.250806] env[61936]: DEBUG nova.compute.manager [None req-d3c79805-05d5-4248-a03b-cbfff298b9ec tempest-ServerDiagnosticsV248Test-1985984208 tempest-ServerDiagnosticsV248Test-1985984208-project-admin] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 587.253683] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb2edd2-92df-4331-83eb-5ba3c84dc27b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.258107] env[61936]: DEBUG nova.compute.utils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.263519] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 587.263763] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 587.269983] env[61936]: INFO nova.compute.manager [None req-d3c79805-05d5-4248-a03b-cbfff298b9ec tempest-ServerDiagnosticsV248Test-1985984208 tempest-ServerDiagnosticsV248Test-1985984208-project-admin] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Retrieving diagnostics [ 587.270873] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a70e0a3-0cc7-4c16-b492-e7cd671fed59 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.349664] env[61936]: DEBUG oslo_concurrency.lockutils [req-6dd9c1a0-be90-4527-8319-b7066ec1c75a req-c214bc71-002e-4dd7-9e20-248fb7b47827 service nova] Releasing lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.351929] env[61936]: DEBUG nova.compute.manager [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Received event network-changed-fa16c095-21c9-4908-b9b7-2b5dfa7aaad5 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 587.352186] env[61936]: DEBUG nova.compute.manager [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Refreshing instance network info cache due to event network-changed-fa16c095-21c9-4908-b9b7-2b5dfa7aaad5. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 587.352319] env[61936]: DEBUG oslo_concurrency.lockutils [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] Acquiring lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.352492] env[61936]: DEBUG oslo_concurrency.lockutils [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] Acquired lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.352598] env[61936]: DEBUG nova.network.neutron [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Refreshing network info cache for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.390822] env[61936]: DEBUG nova.policy [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0d5f48f4bd140b2a5df027dcf90f9bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea52da2527944d0a44ed68dae2d1b45', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.393084] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Releasing lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.393084] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 587.393274] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 587.393549] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc61d629-36c9-4941-aeed-121dabaaba9c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.408606] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3a7cc9-268d-421e-b10b-9b7fa9913c62 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.439471] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ffdb44ea-791a-41b3-9e65-9cf77717ce9a could not be found. [ 587.439987] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.440235] env[61936]: INFO nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 587.440811] env[61936]: DEBUG oslo.service.loopingcall [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.443740] env[61936]: DEBUG nova.compute.manager [-] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 587.443833] env[61936]: DEBUG nova.network.neutron [-] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.482183] env[61936]: ERROR nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. [ 587.482183] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.482183] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 587.482183] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.482183] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.482183] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.482183] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.482183] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.482183] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.482183] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 587.482183] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.482183] env[61936]: ERROR nova.compute.manager raise self.value [ 587.482183] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.482183] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.482183] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.482183] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.482908] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.482908] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.482908] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. [ 587.482908] env[61936]: ERROR nova.compute.manager [ 587.482908] env[61936]: Traceback (most recent call last): [ 587.482908] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.482908] env[61936]: listener.cb(fileno) [ 587.482908] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.482908] env[61936]: result = function(*args, **kwargs) [ 587.482908] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.482908] env[61936]: return func(*args, **kwargs) [ 587.482908] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 587.482908] env[61936]: raise e [ 587.482908] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 587.482908] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 587.482908] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.482908] env[61936]: created_port_ids = self._update_ports_for_instance( [ 587.482908] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.482908] env[61936]: with excutils.save_and_reraise_exception(): [ 587.482908] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.482908] env[61936]: self.force_reraise() [ 587.482908] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.482908] env[61936]: raise self.value [ 587.482908] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.482908] env[61936]: updated_port = self._update_port( [ 587.482908] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.482908] env[61936]: _ensure_no_port_binding_failure(port) [ 587.482908] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.482908] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.483688] env[61936]: nova.exception.PortBindingFailed: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. [ 587.483688] env[61936]: Removing descriptor: 17 [ 587.483688] env[61936]: ERROR nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Traceback (most recent call last): [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] yield resources [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self.driver.spawn(context, instance, image_meta, [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.483688] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] vm_ref = self.build_virtual_machine(instance, [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] for vif in network_info: [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return self._sync_wrapper(fn, *args, **kwargs) [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self.wait() [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self[:] = self._gt.wait() [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return self._exit_event.wait() [ 587.483990] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] result = hub.switch() [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return self.greenlet.switch() [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] result = function(*args, **kwargs) [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return func(*args, **kwargs) [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] raise e [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] nwinfo = self.network_api.allocate_for_instance( [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.484344] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] created_port_ids = self._update_ports_for_instance( [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] with excutils.save_and_reraise_exception(): [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self.force_reraise() [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] raise self.value [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] updated_port = self._update_port( [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] _ensure_no_port_binding_failure(port) [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.484660] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] raise exception.PortBindingFailed(port_id=port['id']) [ 587.484953] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] nova.exception.PortBindingFailed: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. [ 587.484953] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] [ 587.484953] env[61936]: INFO nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Terminating instance [ 587.596797] env[61936]: DEBUG nova.network.neutron [-] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.613295] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11510b17-c959-4dee-bff1-96834d8408eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.622764] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71e3dfd-0853-4c86-9325-2b8a097ced5b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.663687] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5eb897-a31b-48b1-b483-e97a30882c04 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.672409] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b51f410-2855-493f-9561-e41ac6ffeb36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.690554] env[61936]: DEBUG nova.compute.provider_tree [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.768353] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 587.855116] env[61936]: INFO nova.scheduler.client.report [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Deleted allocations for instance bc67f44d-e6c8-49e0-ac2a-cf7449253580 [ 587.929123] env[61936]: DEBUG nova.network.neutron [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.991855] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.100145] env[61936]: DEBUG nova.network.neutron [-] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.138252] env[61936]: DEBUG nova.network.neutron [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.195279] env[61936]: DEBUG nova.scheduler.client.report [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 588.357360] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Successfully created port: ae75a625-8e50-4fa5-94bb-a04853b3b69b {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.369908] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7c68ffde-325e-4952-95c6-da1ca6f15e03 tempest-ServersTestFqdnHostnames-1891333307 tempest-ServersTestFqdnHostnames-1891333307-project-member] Lock "bc67f44d-e6c8-49e0-ac2a-cf7449253580" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.362s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.603759] env[61936]: INFO nova.compute.manager [-] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Took 1.16 seconds to deallocate network for instance. [ 588.606315] env[61936]: DEBUG nova.compute.claims [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.606939] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.640392] env[61936]: DEBUG oslo_concurrency.lockutils [req-f91dbfcc-2310-4a6f-8f6e-235221c39470 req-a6d6f366-585d-4606-a920-e0fd3c9854f1 service nova] Releasing lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.642428] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquired lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.642428] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.701926] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.702684] env[61936]: ERROR nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Traceback (most recent call last): [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self.driver.spawn(context, instance, image_meta, [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] vm_ref = self.build_virtual_machine(instance, [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.702684] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] for vif in network_info: [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return self._sync_wrapper(fn, *args, **kwargs) [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self.wait() [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self[:] = self._gt.wait() [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return self._exit_event.wait() [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] result = hub.switch() [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.703042] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return self.greenlet.switch() [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] result = function(*args, **kwargs) [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] return func(*args, **kwargs) [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] raise e [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] nwinfo = self.network_api.allocate_for_instance( [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] created_port_ids = self._update_ports_for_instance( [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] with excutils.save_and_reraise_exception(): [ 588.703471] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] self.force_reraise() [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] raise self.value [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] updated_port = self._update_port( [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] _ensure_no_port_binding_failure(port) [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] raise exception.PortBindingFailed(port_id=port['id']) [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] nova.exception.PortBindingFailed: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. [ 588.703794] env[61936]: ERROR nova.compute.manager [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] [ 588.704128] env[61936]: DEBUG nova.compute.utils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.706976] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.224s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.706976] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.706976] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 588.706976] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.543s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.707605] env[61936]: INFO nova.compute.claims [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.711079] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Build of instance c4b372a3-f76a-411d-a336-338f5c6d502c was re-scheduled: Binding failed for port dec513a9-aac4-4e9f-ba67-271187b72279, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 588.711669] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 588.711953] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.712156] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquired lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.712430] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.714176] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9578e02e-5058-4f39-ad94-ec972ef1d8d0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.724521] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e48d6c7-282b-4397-9317-55112816d398 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.748312] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08527d50-85cb-415f-bcae-b0977c5e4a36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.752502] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988b948e-7a70-48fb-b91b-6a1b33cdcb48 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.790671] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 588.792918] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181542MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 588.793112] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.876583] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 588.954341] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 588.957613] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.957885] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 588.958183] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.958734] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 588.958734] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 588.958996] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 588.959286] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 588.959579] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 588.960089] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 588.960162] env[61936]: DEBUG nova.virt.hardware [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 588.963346] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239128ca-ed23-49be-a738-0f5d77aeeeba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.971699] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659d8143-810b-4db3-a0e8-f654d1983a7a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.213341] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.272115] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.361201] env[61936]: DEBUG nova.compute.manager [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Received event network-vif-deleted-977a9bb1-444c-48a0-a662-fb19f32b5e72 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 589.361201] env[61936]: DEBUG nova.compute.manager [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Received event network-changed-9fcab735-be3b-48c2-a158-50e091ace076 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 589.361201] env[61936]: DEBUG nova.compute.manager [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Refreshing instance network info cache due to event network-changed-9fcab735-be3b-48c2-a158-50e091ace076. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 589.361201] env[61936]: DEBUG oslo_concurrency.lockutils [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] Acquiring lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.362795] env[61936]: DEBUG oslo_concurrency.lockutils [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] Acquired lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.363848] env[61936]: DEBUG nova.network.neutron [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Refreshing network info cache for port 9fcab735-be3b-48c2-a158-50e091ace076 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 589.411971] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.413275] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.506317] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.917178] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Releasing lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.917382] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 589.917592] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.919377] env[61936]: DEBUG nova.network.neutron [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.921667] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb6b4b4c-351c-47c3-abae-4b319d92977d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.933785] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388282b0-7f02-4533-b256-ac8edaeb1e65 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.966788] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 09467d10-8235-4712-b00f-d74bc6ed39cc could not be found. [ 589.966788] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.966869] env[61936]: INFO nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 589.967906] env[61936]: DEBUG oslo.service.loopingcall [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.967906] env[61936]: DEBUG nova.compute.manager [-] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 589.967906] env[61936]: DEBUG nova.network.neutron [-] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.011134] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Releasing lock "refresh_cache-c4b372a3-f76a-411d-a336-338f5c6d502c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.011134] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 590.011134] env[61936]: DEBUG nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 590.011134] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.019493] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc41b3c-6afb-4421-afcc-3e76725d3ea4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.028041] env[61936]: DEBUG nova.network.neutron [-] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.030088] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c26bba6-1cb8-4982-ad3b-52c6454a3744 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.064184] env[61936]: DEBUG nova.network.neutron [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.065803] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edbcb4b-c266-48b9-a8b1-dfdcf3b1177a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.074653] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.076591] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab82e5c1-9b45-4c5f-a79e-5ab467bc5e54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.092814] env[61936]: DEBUG nova.compute.provider_tree [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.113204] env[61936]: DEBUG nova.compute.manager [req-306f6708-bb01-4cc4-ae3a-056030c15638 req-bbb7a49d-cbcb-4ce7-8d5d-687b3e4d8dde service nova] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Received event network-vif-deleted-fa16c095-21c9-4908-b9b7-2b5dfa7aaad5 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 590.533950] env[61936]: DEBUG nova.network.neutron [-] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.570323] env[61936]: DEBUG oslo_concurrency.lockutils [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] Releasing lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.570420] env[61936]: DEBUG nova.compute.manager [req-9d736372-52bf-4179-90fb-9207502cb48a req-76ceb182-2456-42a7-bed8-80d5e0a06655 service nova] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Received event network-vif-deleted-9fcab735-be3b-48c2-a158-50e091ace076 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 590.581300] env[61936]: DEBUG nova.network.neutron [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.596989] env[61936]: DEBUG nova.scheduler.client.report [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 591.037525] env[61936]: INFO nova.compute.manager [-] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Took 1.07 seconds to deallocate network for instance. [ 591.040347] env[61936]: DEBUG nova.compute.claims [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.040539] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.084459] env[61936]: INFO nova.compute.manager [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: c4b372a3-f76a-411d-a336-338f5c6d502c] Took 1.07 seconds to deallocate network for instance. [ 591.102768] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.104398] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 591.111517] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.538s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.611069] env[61936]: DEBUG nova.compute.utils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.612477] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 591.612646] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.854922] env[61936]: DEBUG nova.policy [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0d5f48f4bd140b2a5df027dcf90f9bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea52da2527944d0a44ed68dae2d1b45', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.914953] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1396a1bc-7fe1-4228-8bcc-695d1aa5b00a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.937542] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04ce8c0-c29f-4f98-a9c5-47ad4dddbfa3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.979396] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9258e4f8-0695-4a70-a5ac-f09dc271a231 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.987816] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70746747-ac0c-49fa-8341-3ffda7682f1c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.002342] env[61936]: DEBUG nova.compute.provider_tree [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.119449] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 592.127667] env[61936]: INFO nova.scheduler.client.report [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Deleted allocations for instance c4b372a3-f76a-411d-a336-338f5c6d502c [ 592.505353] env[61936]: DEBUG nova.scheduler.client.report [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 592.636796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e441d9e4-3ad1-412b-86e4-bb0f7e54d0c5 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "c4b372a3-f76a-411d-a336-338f5c6d502c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.799s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.011598] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.902s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.012352] env[61936]: ERROR nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Traceback (most recent call last): [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self.driver.spawn(context, instance, image_meta, [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] vm_ref = self.build_virtual_machine(instance, [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.012352] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] for vif in network_info: [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return self._sync_wrapper(fn, *args, **kwargs) [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self.wait() [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self[:] = self._gt.wait() [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return self._exit_event.wait() [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] result = hub.switch() [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.012754] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return self.greenlet.switch() [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] result = function(*args, **kwargs) [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] return func(*args, **kwargs) [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] raise e [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] nwinfo = self.network_api.allocate_for_instance( [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] created_port_ids = self._update_ports_for_instance( [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] with excutils.save_and_reraise_exception(): [ 593.013079] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] self.force_reraise() [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] raise self.value [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] updated_port = self._update_port( [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] _ensure_no_port_binding_failure(port) [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] raise exception.PortBindingFailed(port_id=port['id']) [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] nova.exception.PortBindingFailed: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. [ 593.013785] env[61936]: ERROR nova.compute.manager [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] [ 593.017346] env[61936]: DEBUG nova.compute.utils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.017346] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.850s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.019564] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Build of instance 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9 was re-scheduled: Binding failed for port 3c8b50ba-50ab-4feb-bde4-dc4b83184616, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 593.020186] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 593.020514] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Acquiring lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.020771] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Acquired lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.021117] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.129292] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 593.146075] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 593.161399] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 593.161657] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.161820] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 593.162131] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.162216] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 593.162415] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 593.162465] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 593.162615] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 593.162767] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 593.162919] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 593.163091] env[61936]: DEBUG nova.virt.hardware [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 593.164054] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007106c7-072e-4a33-9896-95fac97d28d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.172593] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db3105f-9752-4a4c-9f4b-f0056f7f24cb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.451587] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Acquiring lock "ab8313a4-b427-456b-ab81-19a0b0f95d71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.451844] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Lock "ab8313a4-b427-456b-ab81-19a0b0f95d71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.465084] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Successfully created port: 04a65b8a-bb0f-4483-b806-6f1815813309 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.589733] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.618777] env[61936]: ERROR nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. [ 593.618777] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 593.618777] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 593.618777] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 593.618777] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.618777] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 593.618777] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.618777] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 593.618777] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.618777] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 593.618777] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.618777] env[61936]: ERROR nova.compute.manager raise self.value [ 593.618777] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.618777] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 593.618777] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.618777] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 593.619239] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.619239] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 593.619239] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. [ 593.619239] env[61936]: ERROR nova.compute.manager [ 593.619239] env[61936]: Traceback (most recent call last): [ 593.619239] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 593.619239] env[61936]: listener.cb(fileno) [ 593.619239] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.619239] env[61936]: result = function(*args, **kwargs) [ 593.619239] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.619239] env[61936]: return func(*args, **kwargs) [ 593.619239] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 593.619239] env[61936]: raise e [ 593.619239] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 593.619239] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 593.619239] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.619239] env[61936]: created_port_ids = self._update_ports_for_instance( [ 593.619239] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.619239] env[61936]: with excutils.save_and_reraise_exception(): [ 593.619239] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.619239] env[61936]: self.force_reraise() [ 593.619239] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.619239] env[61936]: raise self.value [ 593.619239] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.619239] env[61936]: updated_port = self._update_port( [ 593.619239] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.619239] env[61936]: _ensure_no_port_binding_failure(port) [ 593.619239] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.619239] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 593.620508] env[61936]: nova.exception.PortBindingFailed: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. [ 593.620508] env[61936]: Removing descriptor: 19 [ 593.620508] env[61936]: ERROR nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Traceback (most recent call last): [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] yield resources [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self.driver.spawn(context, instance, image_meta, [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.620508] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] vm_ref = self.build_virtual_machine(instance, [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] for vif in network_info: [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return self._sync_wrapper(fn, *args, **kwargs) [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self.wait() [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self[:] = self._gt.wait() [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return self._exit_event.wait() [ 593.620868] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] result = hub.switch() [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return self.greenlet.switch() [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] result = function(*args, **kwargs) [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return func(*args, **kwargs) [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] raise e [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] nwinfo = self.network_api.allocate_for_instance( [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.621253] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] created_port_ids = self._update_ports_for_instance( [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] with excutils.save_and_reraise_exception(): [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self.force_reraise() [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] raise self.value [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] updated_port = self._update_port( [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] _ensure_no_port_binding_failure(port) [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.621679] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] raise exception.PortBindingFailed(port_id=port['id']) [ 593.622057] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] nova.exception.PortBindingFailed: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. [ 593.622057] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] [ 593.622057] env[61936]: INFO nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Terminating instance [ 593.685657] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.827257] env[61936]: DEBUG nova.compute.manager [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Received event network-changed-ae75a625-8e50-4fa5-94bb-a04853b3b69b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 593.827257] env[61936]: DEBUG nova.compute.manager [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Refreshing instance network info cache due to event network-changed-ae75a625-8e50-4fa5-94bb-a04853b3b69b. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 593.827257] env[61936]: DEBUG oslo_concurrency.lockutils [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] Acquiring lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.827257] env[61936]: DEBUG oslo_concurrency.lockutils [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] Acquired lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.827257] env[61936]: DEBUG nova.network.neutron [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Refreshing network info cache for port ae75a625-8e50-4fa5-94bb-a04853b3b69b {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 593.875522] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b693c18-45c3-43b0-ad82-b010d66e6600 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.884755] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19f0ca0-c7c0-463f-85e8-22099fbf042b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.890503] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.926954] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedde9a3-82f7-4326-8f81-74b9327b92d9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.935650] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e3f9da-385b-468a-8254-aa1818cb828a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.950550] env[61936]: DEBUG nova.compute.provider_tree [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.129437] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.368075] env[61936]: DEBUG nova.network.neutron [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.393792] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Releasing lock "refresh_cache-449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.394012] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 594.394054] env[61936]: DEBUG nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 594.394446] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.456212] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.459949] env[61936]: DEBUG nova.scheduler.client.report [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 594.753163] env[61936]: DEBUG nova.network.neutron [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.965147] env[61936]: DEBUG nova.network.neutron [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.970071] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.970698] env[61936]: ERROR nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Traceback (most recent call last): [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self.driver.spawn(context, instance, image_meta, [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] vm_ref = self.build_virtual_machine(instance, [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.970698] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] for vif in network_info: [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return self._sync_wrapper(fn, *args, **kwargs) [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self.wait() [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self[:] = self._gt.wait() [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return self._exit_event.wait() [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] result = hub.switch() [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.971115] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return self.greenlet.switch() [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] result = function(*args, **kwargs) [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] return func(*args, **kwargs) [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] raise e [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] nwinfo = self.network_api.allocate_for_instance( [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] created_port_ids = self._update_ports_for_instance( [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] with excutils.save_and_reraise_exception(): [ 594.971462] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] self.force_reraise() [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] raise self.value [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] updated_port = self._update_port( [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] _ensure_no_port_binding_failure(port) [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] raise exception.PortBindingFailed(port_id=port['id']) [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] nova.exception.PortBindingFailed: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. [ 594.971810] env[61936]: ERROR nova.compute.manager [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] [ 594.972110] env[61936]: DEBUG nova.compute.utils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 594.973344] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.243s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.978800] env[61936]: INFO nova.compute.claims [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.982783] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Build of instance 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2 was re-scheduled: Binding failed for port 383074b9-b966-443d-a4fc-519c9725d03c, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 594.988890] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 594.988890] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Acquiring lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.988890] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Acquired lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.988890] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.258296] env[61936]: DEBUG oslo_concurrency.lockutils [req-c6e9288a-4ab0-42a3-a71c-6420bbbc163e req-0922d3d8-7cbf-4932-b358-1065d15ac58d service nova] Releasing lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.258296] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquired lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.258296] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.470254] env[61936]: INFO nova.compute.manager [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] [instance: 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9] Took 1.08 seconds to deallocate network for instance. [ 595.611512] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.797829] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.810266] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.131185] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.220933] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "fac84980-b2ba-4de8-afed-228a28a0f299" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.220933] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "fac84980-b2ba-4de8-afed-228a28a0f299" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.315996] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Releasing lock "refresh_cache-3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.317125] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 596.317398] env[61936]: DEBUG nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 596.317773] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.365982] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb258aa-2362-4d5a-aa6a-aaab4fe44348 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.375422] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9325a0e-00c3-4ed7-b33f-0f5b1962daa6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.380723] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.415109] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6100b8-de81-421b-b8af-6590f0e1e2d3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.427308] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0121fd98-c720-4cc4-be5b-3a67b0072ba8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.447056] env[61936]: DEBUG nova.compute.provider_tree [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.517174] env[61936]: INFO nova.scheduler.client.report [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Deleted allocations for instance 449cec0e-bc1f-4d9d-9e01-eee5eb985ef9 [ 596.634142] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Releasing lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.634576] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 596.634765] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 596.635091] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1b99c88-c050-487a-aa83-695de7d9bb9d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.644358] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e759482-2037-4348-8db8-139ed3aea7ff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.665407] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10 could not be found. [ 596.665624] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 596.665801] env[61936]: INFO nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Took 0.03 seconds to destroy the instance on the hypervisor. [ 596.666093] env[61936]: DEBUG oslo.service.loopingcall [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.666270] env[61936]: DEBUG nova.compute.manager [-] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 596.666364] env[61936]: DEBUG nova.network.neutron [-] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.736145] env[61936]: DEBUG nova.network.neutron [-] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.885776] env[61936]: DEBUG nova.network.neutron [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.957484] env[61936]: DEBUG nova.scheduler.client.report [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 597.027460] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bfb11090-cb2c-46aa-89a2-f736cab5361c tempest-ServerDiagnosticsNegativeTest-1563829579 tempest-ServerDiagnosticsNegativeTest-1563829579-project-member] Lock "449cec0e-bc1f-4d9d-9e01-eee5eb985ef9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.687s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.238885] env[61936]: DEBUG nova.network.neutron [-] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.387738] env[61936]: INFO nova.compute.manager [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] [instance: 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2] Took 1.07 seconds to deallocate network for instance. [ 597.464388] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.465171] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 597.470731] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.417s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.529948] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 597.540503] env[61936]: DEBUG nova.compute.manager [req-b729db8a-7763-4b33-8f0e-83065192ec37 req-abb26347-07a0-4faf-b64a-e2af6ed1d0d9 service nova] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Received event network-vif-deleted-ae75a625-8e50-4fa5-94bb-a04853b3b69b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 597.670090] env[61936]: ERROR nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. [ 597.670090] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.670090] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.670090] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.670090] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.670090] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.670090] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.670090] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.670090] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.670090] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 597.670090] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.670090] env[61936]: ERROR nova.compute.manager raise self.value [ 597.670090] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.670090] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.670090] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.670090] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.670495] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.670495] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.670495] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. [ 597.670495] env[61936]: ERROR nova.compute.manager [ 597.670495] env[61936]: Traceback (most recent call last): [ 597.670495] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.670495] env[61936]: listener.cb(fileno) [ 597.670495] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.670495] env[61936]: result = function(*args, **kwargs) [ 597.670495] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.670495] env[61936]: return func(*args, **kwargs) [ 597.670495] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.670495] env[61936]: raise e [ 597.670495] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.670495] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 597.670495] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.670495] env[61936]: created_port_ids = self._update_ports_for_instance( [ 597.670495] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.670495] env[61936]: with excutils.save_and_reraise_exception(): [ 597.670495] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.670495] env[61936]: self.force_reraise() [ 597.670495] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.670495] env[61936]: raise self.value [ 597.670495] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.670495] env[61936]: updated_port = self._update_port( [ 597.670495] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.670495] env[61936]: _ensure_no_port_binding_failure(port) [ 597.670495] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.670495] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.671345] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. [ 597.671345] env[61936]: Removing descriptor: 15 [ 597.671345] env[61936]: ERROR nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Traceback (most recent call last): [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] yield resources [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self.driver.spawn(context, instance, image_meta, [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.671345] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] vm_ref = self.build_virtual_machine(instance, [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] for vif in network_info: [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return self._sync_wrapper(fn, *args, **kwargs) [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self.wait() [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self[:] = self._gt.wait() [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return self._exit_event.wait() [ 597.671656] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] result = hub.switch() [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return self.greenlet.switch() [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] result = function(*args, **kwargs) [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return func(*args, **kwargs) [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] raise e [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] nwinfo = self.network_api.allocate_for_instance( [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.671966] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] created_port_ids = self._update_ports_for_instance( [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] with excutils.save_and_reraise_exception(): [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self.force_reraise() [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] raise self.value [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] updated_port = self._update_port( [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] _ensure_no_port_binding_failure(port) [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.672347] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] raise exception.PortBindingFailed(port_id=port['id']) [ 597.672645] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] nova.exception.PortBindingFailed: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. [ 597.672645] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] [ 597.672645] env[61936]: INFO nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Terminating instance [ 597.743182] env[61936]: INFO nova.compute.manager [-] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Took 1.08 seconds to deallocate network for instance. [ 597.747241] env[61936]: DEBUG nova.compute.claims [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.747241] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.812229] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Acquiring lock "ecf38466-be6f-432a-bdf4-f2403e53c024" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.812460] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Lock "ecf38466-be6f-432a-bdf4-f2403e53c024" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.976761] env[61936]: DEBUG nova.compute.utils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 597.985272] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 597.986120] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 598.065729] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.106632] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c318f2c7-c052-4e2f-a950-1cdb9054f319 tempest-ServersListShow296Test-359310352 tempest-ServersListShow296Test-359310352-project-member] Acquiring lock "62fe885c-6320-4d4a-86f8-4a5afcc6f928" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.106894] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c318f2c7-c052-4e2f-a950-1cdb9054f319 tempest-ServersListShow296Test-359310352 tempest-ServersListShow296Test-359310352-project-member] Lock "62fe885c-6320-4d4a-86f8-4a5afcc6f928" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.137668] env[61936]: DEBUG nova.policy [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9991feca67204516ab52c5bf6c3e4512', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9ffe5ad26954e128fc650f2cdcebb90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.176126] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.176308] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquired lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.176479] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.320197] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b69d7e-6d16-49b2-a40f-f908e6981059 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.333548] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7e1902-0b03-4d74-88c8-18179efb37ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.373272] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40406d95-d62b-409c-9fa8-a59af90323b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.381536] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d690f3-8144-4e34-90b5-da57932aeac6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.399673] env[61936]: DEBUG nova.compute.provider_tree [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.442946] env[61936]: INFO nova.scheduler.client.report [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Deleted allocations for instance 3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2 [ 598.488437] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 598.553513] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.553752] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.689420] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.690088] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.707080] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.905984] env[61936]: DEBUG nova.scheduler.client.report [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 598.914082] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.953519] env[61936]: DEBUG oslo_concurrency.lockutils [None req-76f31198-c281-4271-9528-db961c50fec3 tempest-ServersAdminNegativeTestJSON-1962508410 tempest-ServersAdminNegativeTestJSON-1962508410-project-member] Lock "3214e2af-1f0e-453d-b4e1-c4c9bd2bb3d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.819s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.997706] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Successfully created port: a9fad4ea-4497-4063-872f-36e6fe643096 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.413015] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.942s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.414478] env[61936]: ERROR nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Traceback (most recent call last): [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self.driver.spawn(context, instance, image_meta, [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] vm_ref = self.build_virtual_machine(instance, [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.414478] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] for vif in network_info: [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return self._sync_wrapper(fn, *args, **kwargs) [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self.wait() [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self[:] = self._gt.wait() [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return self._exit_event.wait() [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] result = hub.switch() [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.414852] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return self.greenlet.switch() [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] result = function(*args, **kwargs) [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] return func(*args, **kwargs) [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] raise e [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] nwinfo = self.network_api.allocate_for_instance( [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] created_port_ids = self._update_ports_for_instance( [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] with excutils.save_and_reraise_exception(): [ 599.415341] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] self.force_reraise() [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] raise self.value [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] updated_port = self._update_port( [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] _ensure_no_port_binding_failure(port) [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] raise exception.PortBindingFailed(port_id=port['id']) [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] nova.exception.PortBindingFailed: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. [ 599.415697] env[61936]: ERROR nova.compute.manager [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] [ 599.416052] env[61936]: DEBUG nova.compute.utils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.416507] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.810s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.420905] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Releasing lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.421632] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 599.421697] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.422714] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Build of instance 5aa38a34-2f1c-4546-9688-980d65dfed27 was re-scheduled: Binding failed for port 977a9bb1-444c-48a0-a662-fb19f32b5e72, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 599.423112] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 599.423906] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Acquiring lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.423906] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Acquired lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.424065] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 599.425557] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9119b5a-051b-4b72-a7f3-e14f62602f58 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.440151] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aedbbd1-ad77-4889-acef-27750f99c8a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.459173] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 599.476080] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1944e478-e7ff-4c20-bdfe-95016613c704 could not be found. [ 599.476289] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.476482] env[61936]: INFO nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Took 0.05 seconds to destroy the instance on the hypervisor. [ 599.476748] env[61936]: DEBUG oslo.service.loopingcall [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.476971] env[61936]: DEBUG nova.compute.manager [-] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 599.477078] env[61936]: DEBUG nova.network.neutron [-] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.502138] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 599.511314] env[61936]: DEBUG nova.network.neutron [-] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.539607] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 599.539861] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.540022] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 599.540258] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.540345] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 599.540517] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 599.540741] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 599.542153] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 599.542153] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 599.542153] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 599.542153] env[61936]: DEBUG nova.virt.hardware [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 599.542774] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e647866e-faf7-4372-b143-22a72854570d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.551625] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35be532-a030-43f5-ae4c-1819b2f47e31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.971171] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.015508] env[61936]: DEBUG nova.network.neutron [-] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.017936] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.311558] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61549dba-bae7-470e-88a0-8fbb25cc06e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.320520] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25dbd7d-a6b3-45cc-b9ad-006e58c67610 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.368032] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cb541e-bc9e-41fd-bf07-203d0f6d51ed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.373375] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.385366] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28a6c5e-9ae2-488c-a20e-7970413cf61d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.413584] env[61936]: DEBUG nova.compute.provider_tree [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.442602] env[61936]: DEBUG nova.compute.manager [None req-6e262329-d9d6-404b-8061-2efd88dae887 tempest-ServerDiagnosticsV248Test-1985984208 tempest-ServerDiagnosticsV248Test-1985984208-project-admin] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 600.443837] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ffc75d-6b9f-4914-ae63-f5cc26f42735 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.451765] env[61936]: INFO nova.compute.manager [None req-6e262329-d9d6-404b-8061-2efd88dae887 tempest-ServerDiagnosticsV248Test-1985984208 tempest-ServerDiagnosticsV248Test-1985984208-project-admin] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Retrieving diagnostics [ 600.452900] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5de971-3204-4287-9280-6753b1e2d8de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.486857] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "64412fc6-956f-43a6-bd5c-55745360e480" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.487111] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "64412fc6-956f-43a6-bd5c-55745360e480" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.519988] env[61936]: INFO nova.compute.manager [-] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Took 1.04 seconds to deallocate network for instance. [ 600.521934] env[61936]: DEBUG nova.compute.claims [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.522142] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.876239] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Releasing lock "refresh_cache-5aa38a34-2f1c-4546-9688-980d65dfed27" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.876490] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 600.876655] env[61936]: DEBUG nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 600.876818] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 600.911477] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.916270] env[61936]: DEBUG nova.scheduler.client.report [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 601.095877] env[61936]: ERROR nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. [ 601.095877] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.095877] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.095877] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.095877] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.095877] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.095877] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.095877] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.095877] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.095877] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 601.095877] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.095877] env[61936]: ERROR nova.compute.manager raise self.value [ 601.095877] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.095877] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.095877] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.095877] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.096809] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.096809] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.096809] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. [ 601.096809] env[61936]: ERROR nova.compute.manager [ 601.096809] env[61936]: Traceback (most recent call last): [ 601.096809] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.096809] env[61936]: listener.cb(fileno) [ 601.096809] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.096809] env[61936]: result = function(*args, **kwargs) [ 601.096809] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.096809] env[61936]: return func(*args, **kwargs) [ 601.096809] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.096809] env[61936]: raise e [ 601.096809] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.096809] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 601.096809] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.096809] env[61936]: created_port_ids = self._update_ports_for_instance( [ 601.096809] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.096809] env[61936]: with excutils.save_and_reraise_exception(): [ 601.096809] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.096809] env[61936]: self.force_reraise() [ 601.096809] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.096809] env[61936]: raise self.value [ 601.096809] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.096809] env[61936]: updated_port = self._update_port( [ 601.096809] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.096809] env[61936]: _ensure_no_port_binding_failure(port) [ 601.096809] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.096809] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.097858] env[61936]: nova.exception.PortBindingFailed: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. [ 601.097858] env[61936]: Removing descriptor: 15 [ 601.097858] env[61936]: ERROR nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Traceback (most recent call last): [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] yield resources [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self.driver.spawn(context, instance, image_meta, [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.097858] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] vm_ref = self.build_virtual_machine(instance, [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] for vif in network_info: [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return self._sync_wrapper(fn, *args, **kwargs) [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self.wait() [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self[:] = self._gt.wait() [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return self._exit_event.wait() [ 601.098233] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] result = hub.switch() [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return self.greenlet.switch() [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] result = function(*args, **kwargs) [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return func(*args, **kwargs) [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] raise e [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] nwinfo = self.network_api.allocate_for_instance( [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.098606] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] created_port_ids = self._update_ports_for_instance( [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] with excutils.save_and_reraise_exception(): [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self.force_reraise() [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] raise self.value [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] updated_port = self._update_port( [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] _ensure_no_port_binding_failure(port) [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.098936] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] raise exception.PortBindingFailed(port_id=port['id']) [ 601.099248] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] nova.exception.PortBindingFailed: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. [ 601.099248] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] [ 601.099248] env[61936]: INFO nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Terminating instance [ 601.415135] env[61936]: DEBUG nova.network.neutron [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.422097] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.423016] env[61936]: ERROR nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Traceback (most recent call last): [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self.driver.spawn(context, instance, image_meta, [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] vm_ref = self.build_virtual_machine(instance, [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.423016] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] for vif in network_info: [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return self._sync_wrapper(fn, *args, **kwargs) [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self.wait() [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self[:] = self._gt.wait() [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return self._exit_event.wait() [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] result = hub.switch() [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.424661] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return self.greenlet.switch() [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] result = function(*args, **kwargs) [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] return func(*args, **kwargs) [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] raise e [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] nwinfo = self.network_api.allocate_for_instance( [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] created_port_ids = self._update_ports_for_instance( [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] with excutils.save_and_reraise_exception(): [ 601.425237] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] self.force_reraise() [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] raise self.value [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] updated_port = self._update_port( [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] _ensure_no_port_binding_failure(port) [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] raise exception.PortBindingFailed(port_id=port['id']) [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] nova.exception.PortBindingFailed: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. [ 601.425600] env[61936]: ERROR nova.compute.manager [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] [ 601.425865] env[61936]: DEBUG nova.compute.utils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.425865] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.632s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.429020] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Build of instance ffdb44ea-791a-41b3-9e65-9cf77717ce9a was re-scheduled: Binding failed for port 9fcab735-be3b-48c2-a158-50e091ace076, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 601.429020] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 601.429020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Acquiring lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.429020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Acquired lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.429711] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.527818] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Acquiring lock "aea12163-da33-436d-ad50-00f8c4fc8ae7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.528077] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Lock "aea12163-da33-436d-ad50-00f8c4fc8ae7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.606045] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Acquiring lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.606045] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Acquired lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.606045] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.850788] env[61936]: DEBUG nova.compute.manager [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Received event network-changed-04a65b8a-bb0f-4483-b806-6f1815813309 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 601.851571] env[61936]: DEBUG nova.compute.manager [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Refreshing instance network info cache due to event network-changed-04a65b8a-bb0f-4483-b806-6f1815813309. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 601.851571] env[61936]: DEBUG oslo_concurrency.lockutils [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] Acquiring lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.851571] env[61936]: DEBUG oslo_concurrency.lockutils [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] Acquired lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.851571] env[61936]: DEBUG nova.network.neutron [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Refreshing network info cache for port 04a65b8a-bb0f-4483-b806-6f1815813309 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 601.917954] env[61936]: INFO nova.compute.manager [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] [instance: 5aa38a34-2f1c-4546-9688-980d65dfed27] Took 1.04 seconds to deallocate network for instance. [ 601.983991] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.073625] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.137567] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.242970] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.403972] env[61936]: DEBUG nova.network.neutron [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.478080] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 7dadef86-7685-40e2-9e68-debc4b4a8a95 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 602.479501] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Acquiring lock "df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.479713] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Lock "df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.576844] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Releasing lock "refresh_cache-ffdb44ea-791a-41b3-9e65-9cf77717ce9a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.577111] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 602.577339] env[61936]: DEBUG nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 602.577523] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.612053] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.644387] env[61936]: DEBUG nova.network.neutron [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.750875] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Releasing lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.751399] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 602.751743] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.752043] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdf2f9f3-3416-4a05-befb-bf5f51461987 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.763535] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b9b885-048f-4173-8abf-923d99392ac5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.788262] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0257bab6-c040-4f1f-8218-c5bc353777bc could not be found. [ 602.788687] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.789033] env[61936]: INFO nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 602.789529] env[61936]: DEBUG oslo.service.loopingcall [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.791690] env[61936]: DEBUG nova.compute.manager [-] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 602.791690] env[61936]: DEBUG nova.network.neutron [-] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.809505] env[61936]: DEBUG nova.network.neutron [-] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.958401] env[61936]: INFO nova.scheduler.client.report [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Deleted allocations for instance 5aa38a34-2f1c-4546-9688-980d65dfed27 [ 602.984698] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 5aa38a34-2f1c-4546-9688-980d65dfed27 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 603.113898] env[61936]: DEBUG nova.network.neutron [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.147939] env[61936]: DEBUG oslo_concurrency.lockutils [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] Releasing lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.148881] env[61936]: DEBUG nova.compute.manager [req-f865d1d5-052b-41ad-9f08-51fd36307c8a req-71fb4f34-316b-4e83-8f9b-0bf154322bb1 service nova] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Received event network-vif-deleted-04a65b8a-bb0f-4483-b806-6f1815813309 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 603.312255] env[61936]: DEBUG nova.network.neutron [-] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.468553] env[61936]: DEBUG oslo_concurrency.lockutils [None req-138c08e6-4aad-4bb7-a06d-a442e6ab644e tempest-ServerExternalEventsTest-1856420833 tempest-ServerExternalEventsTest-1856420833-project-member] Lock "5aa38a34-2f1c-4546-9688-980d65dfed27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.959s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.486022] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance ffdb44ea-791a-41b3-9e65-9cf77717ce9a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 603.486167] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 09467d10-8235-4712-b00f-d74bc6ed39cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 603.486275] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 603.486800] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 1944e478-e7ff-4c20-bdfe-95016613c704 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 603.486800] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 0257bab6-c040-4f1f-8218-c5bc353777bc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 603.499036] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Acquiring lock "c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.499036] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Lock "c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.589832] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "a942e040-549e-4aca-8c4c-67de236e746e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.589832] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "a942e040-549e-4aca-8c4c-67de236e746e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.615904] env[61936]: INFO nova.compute.manager [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] [instance: ffdb44ea-791a-41b3-9e65-9cf77717ce9a] Took 1.04 seconds to deallocate network for instance. [ 603.639628] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "7dadef86-7685-40e2-9e68-debc4b4a8a95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.639628] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "7dadef86-7685-40e2-9e68-debc4b4a8a95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.639628] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "7dadef86-7685-40e2-9e68-debc4b4a8a95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.639628] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "7dadef86-7685-40e2-9e68-debc4b4a8a95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.639839] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "7dadef86-7685-40e2-9e68-debc4b4a8a95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.642071] env[61936]: INFO nova.compute.manager [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Terminating instance [ 603.818340] env[61936]: INFO nova.compute.manager [-] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Took 1.03 seconds to deallocate network for instance. [ 603.821437] env[61936]: DEBUG nova.compute.claims [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.821628] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.971578] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 603.990057] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e3f6ec50-6d1b-46bd-889d-efe70bc7b98f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 604.150114] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "refresh_cache-7dadef86-7685-40e2-9e68-debc4b4a8a95" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.150114] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquired lock "refresh_cache-7dadef86-7685-40e2-9e68-debc4b4a8a95" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.150114] env[61936]: DEBUG nova.network.neutron [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.493950] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 39a94d7e-77bb-4083-81c3-4dce3bf55fdd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 604.504871] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.598106] env[61936]: DEBUG nova.compute.manager [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Received event network-changed-a9fad4ea-4497-4063-872f-36e6fe643096 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 604.598106] env[61936]: DEBUG nova.compute.manager [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Refreshing instance network info cache due to event network-changed-a9fad4ea-4497-4063-872f-36e6fe643096. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 604.598106] env[61936]: DEBUG oslo_concurrency.lockutils [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] Acquiring lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.598106] env[61936]: DEBUG oslo_concurrency.lockutils [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] Acquired lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.598106] env[61936]: DEBUG nova.network.neutron [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Refreshing network info cache for port a9fad4ea-4497-4063-872f-36e6fe643096 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 604.656692] env[61936]: INFO nova.scheduler.client.report [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Deleted allocations for instance ffdb44ea-791a-41b3-9e65-9cf77717ce9a [ 604.670844] env[61936]: DEBUG nova.network.neutron [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.820306] env[61936]: DEBUG nova.network.neutron [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.821946] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Acquiring lock "ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.822126] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Lock "ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.998536] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance aad2d5db-9e34-40d3-816c-902eecd0069e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 605.166659] env[61936]: DEBUG oslo_concurrency.lockutils [None req-760044a6-b2ce-491d-a8cd-5b1a6843a4a7 tempest-ServerDiagnosticsTest-2087849872 tempest-ServerDiagnosticsTest-2087849872-project-member] Lock "ffdb44ea-791a-41b3-9e65-9cf77717ce9a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.047s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.231179] env[61936]: DEBUG nova.network.neutron [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.323773] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Releasing lock "refresh_cache-7dadef86-7685-40e2-9e68-debc4b4a8a95" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.324665] env[61936]: DEBUG nova.compute.manager [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 605.324862] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.325894] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1c5c5f-177f-4a8f-85d1-a31050114a2e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.338965] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 605.340090] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02c936c2-9312-4ebb-9b19-c16a450e925b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.349035] env[61936]: DEBUG oslo_vmware.api [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 605.349035] env[61936]: value = "task-1252703" [ 605.349035] env[61936]: _type = "Task" [ 605.349035] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.366377] env[61936]: DEBUG oslo_vmware.api [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.498952] env[61936]: DEBUG nova.network.neutron [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.505813] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance de5277a7-05e2-49eb-b577-87f900eeef3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 605.669561] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 605.863701] env[61936]: DEBUG oslo_vmware.api [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252703, 'name': PowerOffVM_Task, 'duration_secs': 0.125481} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.863701] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 605.863701] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 605.865135] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-877bafed-ce51-44fe-8f4a-843b93a80810 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.891691] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 605.891691] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 605.891691] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Deleting the datastore file [datastore1] 7dadef86-7685-40e2-9e68-debc4b4a8a95 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 605.891691] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cd10267-9880-4a32-99bf-baf10c21981f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.897492] env[61936]: DEBUG oslo_vmware.api [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for the task: (returnval){ [ 605.897492] env[61936]: value = "task-1252705" [ 605.897492] env[61936]: _type = "Task" [ 605.897492] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.912549] env[61936]: DEBUG oslo_vmware.api [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.003207] env[61936]: DEBUG oslo_concurrency.lockutils [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] Releasing lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.004920] env[61936]: DEBUG nova.compute.manager [req-cc747da1-ccb0-4a1e-92d1-874efe3c747c req-005c0628-4abc-4437-8b21-daaceb9e1b25 service nova] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Received event network-vif-deleted-a9fad4ea-4497-4063-872f-36e6fe643096 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 606.012126] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance c9302959-c96c-47cb-980d-13493aa185f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 606.206676] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.417976] env[61936]: DEBUG oslo_vmware.api [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Task: {'id': task-1252705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102321} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.418244] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 606.418419] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 606.418579] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.420044] env[61936]: INFO nova.compute.manager [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Took 1.09 seconds to destroy the instance on the hypervisor. [ 606.420044] env[61936]: DEBUG oslo.service.loopingcall [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.420044] env[61936]: DEBUG nova.compute.manager [-] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 606.420044] env[61936]: DEBUG nova.network.neutron [-] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.447730] env[61936]: DEBUG nova.network.neutron [-] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.516069] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 69563e1e-301a-4a7b-8fe4-7f1846478f56 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 606.953013] env[61936]: DEBUG nova.network.neutron [-] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.026278] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 9ac0103e-8424-4518-8acb-343267deb3ab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 607.458261] env[61936]: INFO nova.compute.manager [-] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Took 1.04 seconds to deallocate network for instance. [ 607.528716] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance ab8313a4-b427-456b-ab81-19a0b0f95d71 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 607.968357] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.032439] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fac84980-b2ba-4de8-afed-228a28a0f299 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 608.536522] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance ecf38466-be6f-432a-bdf4-f2403e53c024 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 609.042157] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 62fe885c-6320-4d4a-86f8-4a5afcc6f928 has been deleted (perhaps locally). Deleting allocations that remained for this instance against this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1750}} [ 609.058589] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Cannot delete allocation for 62fe885c-6320-4d4a-86f8-4a5afcc6f928 consumer in placement as consumer does not exist {{(pid=61936) delete_allocation_for_instance /opt/stack/nova/nova/scheduler/client/report.py:2203}} [ 609.561941] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 994ff54f-a26b-40e0-a2c7-502f3598d9bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 610.067099] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 610.141828] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "1e0dade4-ea54-457d-8398-54ba810a83c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.142110] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "1e0dade4-ea54-457d-8398-54ba810a83c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.573091] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 64412fc6-956f-43a6-bd5c-55745360e480 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 610.623328] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Acquiring lock "70e751e6-4109-41fd-a80f-82d22c34ea64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.623328] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Lock "70e751e6-4109-41fd-a80f-82d22c34ea64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.076087] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance aea12163-da33-436d-ad50-00f8c4fc8ae7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 611.580979] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 611.580979] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 611.580979] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 611.996016] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9efd64-f0d5-43f0-b538-d571f41df198 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.005701] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab8db97-79b1-493e-81f4-0db1c1786eff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.040955] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce43829e-adc9-48f8-aa52-d37c981f946e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.055021] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7087943-b86c-4e66-bb22-b160d456fafd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.065787] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.568872] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 613.079956] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 613.080410] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.656s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.080603] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.669s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.082125] env[61936]: INFO nova.compute.claims [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.469698] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec5bd78-cf09-4ad0-aa2f-165dbcb384ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.480169] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f66b06-b892-4604-80a8-fdead56cfb5b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.513745] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34949596-4835-4315-b79d-7fd3e6e588fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.523397] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521c258f-a38d-4154-8c98-cf575c103eb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.535408] env[61936]: DEBUG nova.compute.provider_tree [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.038185] env[61936]: DEBUG nova.scheduler.client.report [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 615.544298] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.544825] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 615.550916] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.510s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.059067] env[61936]: DEBUG nova.compute.utils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.063547] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 616.063738] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 616.164158] env[61936]: DEBUG nova.policy [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0d5f48f4bd140b2a5df027dcf90f9bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea52da2527944d0a44ed68dae2d1b45', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.467692] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48377a6-c797-45b5-8d2c-b46ce9739be4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.476340] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bac68d-4542-4862-a1ac-7e4ae9f098de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.510524] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fdcc7e5-9ac2-4e59-9db4-9864eefa23fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.518513] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42487a86-312e-4d02-947a-230d305a9d2a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.531749] env[61936]: DEBUG nova.compute.provider_tree [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.564036] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 616.714905] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Acquiring lock "64e78c61-85f4-4914-8fab-3ced64335275" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.714905] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Lock "64e78c61-85f4-4914-8fab-3ced64335275" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.816838] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Successfully created port: 732af640-3bab-48bb-a44d-efc86f420c9a {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.034883] env[61936]: DEBUG nova.scheduler.client.report [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 617.544035] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.544035] env[61936]: ERROR nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. [ 617.544035] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Traceback (most recent call last): [ 617.544035] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 617.544035] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self.driver.spawn(context, instance, image_meta, [ 617.544035] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 617.544035] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.544035] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.544035] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] vm_ref = self.build_virtual_machine(instance, [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] for vif in network_info: [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return self._sync_wrapper(fn, *args, **kwargs) [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self.wait() [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self[:] = self._gt.wait() [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return self._exit_event.wait() [ 617.544551] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] result = hub.switch() [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return self.greenlet.switch() [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] result = function(*args, **kwargs) [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] return func(*args, **kwargs) [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] raise e [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] nwinfo = self.network_api.allocate_for_instance( [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.544894] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] created_port_ids = self._update_ports_for_instance( [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] with excutils.save_and_reraise_exception(): [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] self.force_reraise() [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] raise self.value [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] updated_port = self._update_port( [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] _ensure_no_port_binding_failure(port) [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.545293] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] raise exception.PortBindingFailed(port_id=port['id']) [ 617.545665] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] nova.exception.PortBindingFailed: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. [ 617.545665] env[61936]: ERROR nova.compute.manager [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] [ 617.545665] env[61936]: DEBUG nova.compute.utils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 617.545665] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Build of instance 09467d10-8235-4712-b00f-d74bc6ed39cc was re-scheduled: Binding failed for port fa16c095-21c9-4908-b9b7-2b5dfa7aaad5, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 617.545665] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 617.545869] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.545869] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquired lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.545869] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.550612] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.865s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.552622] env[61936]: INFO nova.compute.claims [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.581393] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 617.615826] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 617.616079] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.616729] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 617.616729] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.616729] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 617.616729] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 617.616890] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 617.617158] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 617.617158] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 617.617379] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 617.617499] env[61936]: DEBUG nova.virt.hardware [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 617.618527] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87afcac-09c8-41a3-92d6-8d0393e1f084 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.632285] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d94fe0b-7390-48f8-998b-56146718a9c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.071191] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.152570] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.439762] env[61936]: ERROR nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. [ 618.439762] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.439762] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.439762] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.439762] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.439762] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.439762] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.439762] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.439762] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.439762] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 618.439762] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.439762] env[61936]: ERROR nova.compute.manager raise self.value [ 618.439762] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.439762] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.439762] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.439762] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.440318] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.440318] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.440318] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. [ 618.440318] env[61936]: ERROR nova.compute.manager [ 618.440318] env[61936]: Traceback (most recent call last): [ 618.440318] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.440318] env[61936]: listener.cb(fileno) [ 618.440318] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.440318] env[61936]: result = function(*args, **kwargs) [ 618.440318] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.440318] env[61936]: return func(*args, **kwargs) [ 618.440318] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 618.440318] env[61936]: raise e [ 618.440318] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.440318] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 618.440318] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.440318] env[61936]: created_port_ids = self._update_ports_for_instance( [ 618.440318] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.440318] env[61936]: with excutils.save_and_reraise_exception(): [ 618.440318] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.440318] env[61936]: self.force_reraise() [ 618.440318] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.440318] env[61936]: raise self.value [ 618.440318] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.440318] env[61936]: updated_port = self._update_port( [ 618.440318] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.440318] env[61936]: _ensure_no_port_binding_failure(port) [ 618.440318] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.440318] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.441438] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. [ 618.441438] env[61936]: Removing descriptor: 15 [ 618.441438] env[61936]: ERROR nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Traceback (most recent call last): [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] yield resources [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self.driver.spawn(context, instance, image_meta, [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.441438] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] vm_ref = self.build_virtual_machine(instance, [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] for vif in network_info: [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return self._sync_wrapper(fn, *args, **kwargs) [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self.wait() [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self[:] = self._gt.wait() [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return self._exit_event.wait() [ 618.442071] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] result = hub.switch() [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return self.greenlet.switch() [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] result = function(*args, **kwargs) [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return func(*args, **kwargs) [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] raise e [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] nwinfo = self.network_api.allocate_for_instance( [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.442533] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] created_port_ids = self._update_ports_for_instance( [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] with excutils.save_and_reraise_exception(): [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self.force_reraise() [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] raise self.value [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] updated_port = self._update_port( [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] _ensure_no_port_binding_failure(port) [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.442948] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] raise exception.PortBindingFailed(port_id=port['id']) [ 618.445215] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] nova.exception.PortBindingFailed: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. [ 618.445215] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] [ 618.445215] env[61936]: INFO nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Terminating instance [ 618.571202] env[61936]: DEBUG nova.compute.manager [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Received event network-changed-732af640-3bab-48bb-a44d-efc86f420c9a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 618.571458] env[61936]: DEBUG nova.compute.manager [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Refreshing instance network info cache due to event network-changed-732af640-3bab-48bb-a44d-efc86f420c9a. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 618.572095] env[61936]: DEBUG oslo_concurrency.lockutils [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] Acquiring lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.572095] env[61936]: DEBUG oslo_concurrency.lockutils [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] Acquired lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.572095] env[61936]: DEBUG nova.network.neutron [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Refreshing network info cache for port 732af640-3bab-48bb-a44d-efc86f420c9a {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.656808] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Releasing lock "refresh_cache-09467d10-8235-4712-b00f-d74bc6ed39cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.656808] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 618.656808] env[61936]: DEBUG nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 618.656808] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.684493] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.951305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.985755] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe40fcc4-9518-4b5d-a54f-986a9b9f9c3d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.004206] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51affe6f-83d9-4112-87a2-d29e00efc518 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.041018] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e83367-125b-4f51-802a-61f6181531f0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.049541] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1c3ff2-13ec-4d96-9e24-db3171098c4d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.064045] env[61936]: DEBUG nova.compute.provider_tree [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.103037] env[61936]: DEBUG nova.network.neutron [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.187852] env[61936]: DEBUG nova.network.neutron [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.223140] env[61936]: DEBUG nova.network.neutron [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.568474] env[61936]: DEBUG nova.scheduler.client.report [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 619.695207] env[61936]: INFO nova.compute.manager [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 09467d10-8235-4712-b00f-d74bc6ed39cc] Took 1.03 seconds to deallocate network for instance. [ 619.732615] env[61936]: DEBUG oslo_concurrency.lockutils [req-d7f16bc6-665e-43ad-aa75-8780fb9fbbd4 req-a123e23a-c88b-4aca-9482-95505882dca7 service nova] Releasing lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.732839] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquired lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.733561] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.077015] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.077015] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 620.080248] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.333s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.280149] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.480754] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.587232] env[61936]: DEBUG nova.compute.utils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 620.597719] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 620.598767] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 620.686901] env[61936]: DEBUG nova.policy [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1209ddc535b941839e47bb9434b8e5d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc37dc35db2846ef9a68b0d2f84ccbf4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 620.738778] env[61936]: INFO nova.scheduler.client.report [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Deleted allocations for instance 09467d10-8235-4712-b00f-d74bc6ed39cc [ 620.987485] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Releasing lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.988684] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 620.988966] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.989514] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a1435f1-d5d2-4aec-a349-f2a88b7086ed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.002202] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6763e9b-de4d-4a8a-942e-582c04565f75 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.032994] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e3f6ec50-6d1b-46bd-889d-efe70bc7b98f could not be found. [ 621.033293] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.035578] env[61936]: INFO nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 621.035578] env[61936]: DEBUG oslo.service.loopingcall [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.035578] env[61936]: DEBUG nova.compute.manager [-] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 621.035578] env[61936]: DEBUG nova.network.neutron [-] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.037398] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10900464-3727-41ee-9581-307bd0a31127 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.050295] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075620c8-0170-46a1-99ad-8896aef1e367 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.084231] env[61936]: DEBUG nova.network.neutron [-] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.087918] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05486b1b-9361-4320-bac2-b999be7c25bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.097010] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c6b4ab-544b-4bcb-bce9-0e1941890f48 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.102638] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 621.120736] env[61936]: DEBUG nova.compute.provider_tree [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.257540] env[61936]: DEBUG oslo_concurrency.lockutils [None req-85f4f4b5-1e5a-42e6-ac8c-8a48a443a1f3 tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "09467d10-8235-4712-b00f-d74bc6ed39cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.048s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.261208] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Successfully created port: 5fe3de13-f680-43f2-bed5-24cd390263a0 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.389721] env[61936]: DEBUG nova.compute.manager [req-46801327-7cde-4c4e-8098-d0d9d3b0896d req-4435cd55-aa90-43e2-9378-1aca51b2038f service nova] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Received event network-vif-deleted-732af640-3bab-48bb-a44d-efc86f420c9a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 621.587646] env[61936]: DEBUG nova.network.neutron [-] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.627591] env[61936]: DEBUG nova.scheduler.client.report [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 621.766781] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 622.090071] env[61936]: INFO nova.compute.manager [-] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Took 1.05 seconds to deallocate network for instance. [ 622.092994] env[61936]: DEBUG nova.compute.claims [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 622.093093] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.115120] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 622.136302] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.136589] env[61936]: ERROR nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Traceback (most recent call last): [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self.driver.spawn(context, instance, image_meta, [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] vm_ref = self.build_virtual_machine(instance, [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.136589] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] for vif in network_info: [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return self._sync_wrapper(fn, *args, **kwargs) [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self.wait() [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self[:] = self._gt.wait() [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return self._exit_event.wait() [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] result = hub.switch() [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.136961] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return self.greenlet.switch() [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] result = function(*args, **kwargs) [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] return func(*args, **kwargs) [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] raise e [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] nwinfo = self.network_api.allocate_for_instance( [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] created_port_ids = self._update_ports_for_instance( [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] with excutils.save_and_reraise_exception(): [ 622.137792] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] self.force_reraise() [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] raise self.value [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] updated_port = self._update_port( [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] _ensure_no_port_binding_failure(port) [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] raise exception.PortBindingFailed(port_id=port['id']) [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] nova.exception.PortBindingFailed: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. [ 622.138330] env[61936]: ERROR nova.compute.manager [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] [ 622.138665] env[61936]: DEBUG nova.compute.utils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.145468] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.081s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.151024] env[61936]: INFO nova.compute.claims [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.151024] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Build of instance 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10 was re-scheduled: Binding failed for port ae75a625-8e50-4fa5-94bb-a04853b3b69b, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 622.151279] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 622.151734] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.151911] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquired lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.152086] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.167070] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 622.168351] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.168561] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 622.169568] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.169568] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 622.169568] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 622.170204] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 622.170411] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 622.170584] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 622.170745] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 622.171142] env[61936]: DEBUG nova.virt.hardware [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 622.172078] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b881f070-5138-4f09-b060-c5c41f92acae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.183951] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928b36fe-af42-42df-9415-ed04092d94de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.296213] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.715969] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.158598] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.545427] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d41f41-2982-4395-a3ca-4e74efb1d00d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.554514] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4586276e-5e2a-4c4e-8075-4f8f2d90e10c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.585031] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19a8c86-63f5-4298-9636-90722b5de34a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.595699] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a495ab4-8fa1-4952-89c6-0a2d3159fc10 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.608821] env[61936]: DEBUG nova.compute.provider_tree [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.665055] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Releasing lock "refresh_cache-46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.665846] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 623.665846] env[61936]: DEBUG nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 623.665846] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.702215] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.020031] env[61936]: ERROR nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. [ 624.020031] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.020031] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.020031] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.020031] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.020031] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.020031] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.020031] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.020031] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.020031] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 624.020031] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.020031] env[61936]: ERROR nova.compute.manager raise self.value [ 624.020031] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.020031] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.020031] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.020031] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.020801] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.020801] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.020801] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. [ 624.020801] env[61936]: ERROR nova.compute.manager [ 624.020801] env[61936]: Traceback (most recent call last): [ 624.020801] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.020801] env[61936]: listener.cb(fileno) [ 624.020801] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.020801] env[61936]: result = function(*args, **kwargs) [ 624.020801] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.020801] env[61936]: return func(*args, **kwargs) [ 624.020801] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 624.020801] env[61936]: raise e [ 624.020801] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.020801] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 624.020801] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.020801] env[61936]: created_port_ids = self._update_ports_for_instance( [ 624.020801] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.020801] env[61936]: with excutils.save_and_reraise_exception(): [ 624.020801] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.020801] env[61936]: self.force_reraise() [ 624.020801] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.020801] env[61936]: raise self.value [ 624.020801] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.020801] env[61936]: updated_port = self._update_port( [ 624.020801] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.020801] env[61936]: _ensure_no_port_binding_failure(port) [ 624.020801] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.020801] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.021678] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. [ 624.021678] env[61936]: Removing descriptor: 15 [ 624.021678] env[61936]: ERROR nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Traceback (most recent call last): [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] yield resources [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self.driver.spawn(context, instance, image_meta, [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.021678] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] vm_ref = self.build_virtual_machine(instance, [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] for vif in network_info: [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return self._sync_wrapper(fn, *args, **kwargs) [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self.wait() [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self[:] = self._gt.wait() [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return self._exit_event.wait() [ 624.022068] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] result = hub.switch() [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return self.greenlet.switch() [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] result = function(*args, **kwargs) [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return func(*args, **kwargs) [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] raise e [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] nwinfo = self.network_api.allocate_for_instance( [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.022468] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] created_port_ids = self._update_ports_for_instance( [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] with excutils.save_and_reraise_exception(): [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self.force_reraise() [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] raise self.value [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] updated_port = self._update_port( [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] _ensure_no_port_binding_failure(port) [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.022860] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] raise exception.PortBindingFailed(port_id=port['id']) [ 624.023228] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] nova.exception.PortBindingFailed: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. [ 624.023228] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] [ 624.023228] env[61936]: INFO nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Terminating instance [ 624.114682] env[61936]: DEBUG nova.scheduler.client.report [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 624.206664] env[61936]: DEBUG nova.network.neutron [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.527057] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Acquiring lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.527479] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Acquired lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.527564] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.625133] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.625701] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 624.630802] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.613s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.632390] env[61936]: INFO nova.compute.claims [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.710219] env[61936]: INFO nova.compute.manager [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10] Took 1.04 seconds to deallocate network for instance. [ 624.865062] env[61936]: DEBUG nova.compute.manager [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Received event network-changed-5fe3de13-f680-43f2-bed5-24cd390263a0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 624.865396] env[61936]: DEBUG nova.compute.manager [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Refreshing instance network info cache due to event network-changed-5fe3de13-f680-43f2-bed5-24cd390263a0. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 624.865465] env[61936]: DEBUG oslo_concurrency.lockutils [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] Acquiring lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.060264] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.138965] env[61936]: DEBUG nova.compute.utils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.151368] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 625.152665] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.270454] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.288136] env[61936]: DEBUG nova.policy [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5fe2328df734cea9dca08978fba8134', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edc02335db4d4f60a9ed1235dacdb348', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.425237] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Acquiring lock "9d805bf8-f2cd-4108-9007-2de18e1bf283" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.425571] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Lock "9d805bf8-f2cd-4108-9007-2de18e1bf283" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.651566] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 625.757943] env[61936]: INFO nova.scheduler.client.report [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Deleted allocations for instance 46ff89c1-36e2-4ac1-b8ae-d52dd167bd10 [ 625.772610] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Releasing lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.773031] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 625.773567] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.773567] env[61936]: DEBUG oslo_concurrency.lockutils [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] Acquired lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.773725] env[61936]: DEBUG nova.network.neutron [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Refreshing network info cache for port 5fe3de13-f680-43f2-bed5-24cd390263a0 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.774550] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a9885e8-aaf0-4d0a-ae1a-5d30c149f7ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.788811] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5a8f18-3cb6-4dff-a609-015bb52e44c4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.816487] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39a94d7e-77bb-4083-81c3-4dce3bf55fdd could not be found. [ 625.816487] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.816487] env[61936]: INFO nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 625.816795] env[61936]: DEBUG oslo.service.loopingcall [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.819300] env[61936]: DEBUG nova.compute.manager [-] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 625.820505] env[61936]: DEBUG nova.network.neutron [-] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.867554] env[61936]: DEBUG nova.network.neutron [-] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.157363] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea6ea23-e8ee-42ca-b69d-70f9b92034ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.171023] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca1ca2d-6ae8-40dc-82d5-59d6c9f7df36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.209568] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ee17c2-2788-4ab6-af00-473135e5672d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.217542] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1ee213-6364-4bc3-83dc-bedd527061ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.233962] env[61936]: DEBUG nova.compute.provider_tree [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.272970] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db971bb8-02f6-4828-8079-92b1aae2c765 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "46ff89c1-36e2-4ac1-b8ae-d52dd167bd10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.519s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.328755] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Successfully created port: 3799ef5b-d8b0-44b9-8ee9-372b8eb07303 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.339339] env[61936]: DEBUG nova.network.neutron [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.369272] env[61936]: DEBUG nova.network.neutron [-] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.677193] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 626.722699] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 626.722833] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.722873] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 626.723194] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.723194] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 626.723338] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 626.723523] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 626.723681] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 626.723836] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 626.724240] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 626.724240] env[61936]: DEBUG nova.virt.hardware [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 626.725106] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6425ca-f529-4e6c-b87b-e24bfb5c1098 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.735185] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee7f44f-044c-4503-ad8e-14c7b4f014f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.739781] env[61936]: DEBUG nova.network.neutron [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.742042] env[61936]: DEBUG nova.scheduler.client.report [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 626.778738] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 626.875233] env[61936]: INFO nova.compute.manager [-] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Took 1.05 seconds to deallocate network for instance. [ 626.877099] env[61936]: DEBUG nova.compute.claims [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.877480] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.250156] env[61936]: DEBUG oslo_concurrency.lockutils [req-5bbeec2d-9b05-451c-83ac-09fe0a26c04e req-1dbf1de7-0da7-45c5-8910-0b4887a406af service nova] Releasing lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.250156] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.250156] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 627.252517] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.730s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.327665] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.746462] env[61936]: DEBUG nova.compute.manager [req-99e4718c-75b3-41bc-94b4-18b3804b1b1b req-94119e64-0e2c-4f90-a064-436b4dc44bd3 service nova] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Received event network-vif-deleted-5fe3de13-f680-43f2-bed5-24cd390263a0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 627.762288] env[61936]: DEBUG nova.compute.utils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.771913] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 627.771913] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.938108] env[61936]: DEBUG nova.policy [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '39a626c0384c4f7494d586732204c372', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edf081a49b1d4f1b83170bdd53423e57', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.120091] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b11cd00-816a-44bc-9124-c81d2a1d66a1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.127603] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88811175-b9a2-4da4-875c-5faa8fad8bed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.157640] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f514dcf-3ec9-4306-8463-1cafdae923d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.164903] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd463bf-200a-4d21-9ffb-4a9412af11cc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.177894] env[61936]: DEBUG nova.compute.provider_tree [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.269636] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 628.683249] env[61936]: DEBUG nova.scheduler.client.report [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.058038] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "66e52362-2f67-4e5a-823a-eed7a87d1a1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.058250] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "66e52362-2f67-4e5a-823a-eed7a87d1a1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.072369] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Successfully created port: 8bd92dbc-402e-430d-935a-9732f71bcfe9 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.188741] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.189599] env[61936]: ERROR nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Traceback (most recent call last): [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self.driver.spawn(context, instance, image_meta, [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] vm_ref = self.build_virtual_machine(instance, [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.189599] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] for vif in network_info: [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return self._sync_wrapper(fn, *args, **kwargs) [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self.wait() [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self[:] = self._gt.wait() [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return self._exit_event.wait() [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] result = hub.switch() [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.190768] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return self.greenlet.switch() [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] result = function(*args, **kwargs) [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] return func(*args, **kwargs) [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] raise e [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] nwinfo = self.network_api.allocate_for_instance( [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] created_port_ids = self._update_ports_for_instance( [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] with excutils.save_and_reraise_exception(): [ 629.192680] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] self.force_reraise() [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] raise self.value [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] updated_port = self._update_port( [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] _ensure_no_port_binding_failure(port) [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] raise exception.PortBindingFailed(port_id=port['id']) [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] nova.exception.PortBindingFailed: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. [ 629.193058] env[61936]: ERROR nova.compute.manager [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] [ 629.193388] env[61936]: DEBUG nova.compute.utils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.193388] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.371s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.197893] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Build of instance 1944e478-e7ff-4c20-bdfe-95016613c704 was re-scheduled: Binding failed for port 04a65b8a-bb0f-4483-b806-6f1815813309, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 629.198767] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 629.199240] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.200459] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquired lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.200459] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.286896] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 629.321858] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:43:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='295920031',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1804738923',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 629.322921] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.323232] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 629.323533] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.323694] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 629.323878] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 629.324909] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 629.324909] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 629.324909] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 629.324909] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 629.325123] env[61936]: DEBUG nova.virt.hardware [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 629.329417] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a9b128-5bf7-4b39-8854-53344d97a8b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.335551] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be99dec-2920-4722-9bca-6296129aea77 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.419146] env[61936]: ERROR nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. [ 629.419146] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.419146] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.419146] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.419146] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.419146] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.419146] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.419146] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.419146] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.419146] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 629.419146] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.419146] env[61936]: ERROR nova.compute.manager raise self.value [ 629.419146] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.419146] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.419146] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.419146] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.419699] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.419699] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.419699] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. [ 629.419699] env[61936]: ERROR nova.compute.manager [ 629.419699] env[61936]: Traceback (most recent call last): [ 629.419699] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.419699] env[61936]: listener.cb(fileno) [ 629.419699] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.419699] env[61936]: result = function(*args, **kwargs) [ 629.419699] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.419699] env[61936]: return func(*args, **kwargs) [ 629.419699] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.419699] env[61936]: raise e [ 629.419699] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.419699] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 629.419699] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.419699] env[61936]: created_port_ids = self._update_ports_for_instance( [ 629.419699] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.419699] env[61936]: with excutils.save_and_reraise_exception(): [ 629.419699] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.419699] env[61936]: self.force_reraise() [ 629.419699] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.419699] env[61936]: raise self.value [ 629.419699] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.419699] env[61936]: updated_port = self._update_port( [ 629.419699] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.419699] env[61936]: _ensure_no_port_binding_failure(port) [ 629.419699] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.419699] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.420601] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. [ 629.420601] env[61936]: Removing descriptor: 19 [ 629.420601] env[61936]: ERROR nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Traceback (most recent call last): [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] yield resources [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self.driver.spawn(context, instance, image_meta, [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.420601] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] vm_ref = self.build_virtual_machine(instance, [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] for vif in network_info: [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return self._sync_wrapper(fn, *args, **kwargs) [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self.wait() [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self[:] = self._gt.wait() [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return self._exit_event.wait() [ 629.420941] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] result = hub.switch() [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return self.greenlet.switch() [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] result = function(*args, **kwargs) [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return func(*args, **kwargs) [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] raise e [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] nwinfo = self.network_api.allocate_for_instance( [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.421345] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] created_port_ids = self._update_ports_for_instance( [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] with excutils.save_and_reraise_exception(): [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self.force_reraise() [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] raise self.value [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] updated_port = self._update_port( [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] _ensure_no_port_binding_failure(port) [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.421711] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] raise exception.PortBindingFailed(port_id=port['id']) [ 629.422103] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] nova.exception.PortBindingFailed: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. [ 629.422103] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] [ 629.422103] env[61936]: INFO nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Terminating instance [ 629.733820] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.929017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.929232] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquired lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.929406] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.062401] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.158696] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b74c2d0-1a31-4e62-9323-4770c6a21067 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.166555] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ae97f2-1017-4e67-8339-2ff93b5bdefa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.203188] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4737dc50-61f8-4134-adb7-2b782a898bdd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.211556] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fdb3cf6-6831-48db-bab6-e2dd2eee733b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.228804] env[61936]: DEBUG nova.compute.provider_tree [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.231842] env[61936]: DEBUG nova.compute.manager [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Received event network-changed-3799ef5b-d8b0-44b9-8ee9-372b8eb07303 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 630.235891] env[61936]: DEBUG nova.compute.manager [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Refreshing instance network info cache due to event network-changed-3799ef5b-d8b0-44b9-8ee9-372b8eb07303. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 630.235891] env[61936]: DEBUG oslo_concurrency.lockutils [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] Acquiring lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.472078] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.568070] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Releasing lock "refresh_cache-1944e478-e7ff-4c20-bdfe-95016613c704" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.568070] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 630.568070] env[61936]: DEBUG nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 630.568070] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.705235] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.736358] env[61936]: DEBUG nova.scheduler.client.report [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 630.749781] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.122462] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "a2bb2a06-0fb8-422c-bd16-8f369e397dd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.122700] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "a2bb2a06-0fb8-422c-bd16-8f369e397dd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.207856] env[61936]: DEBUG nova.network.neutron [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.251564] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.251564] env[61936]: ERROR nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. [ 631.251564] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Traceback (most recent call last): [ 631.251564] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 631.251564] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self.driver.spawn(context, instance, image_meta, [ 631.251564] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 631.251564] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.251564] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.251564] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] vm_ref = self.build_virtual_machine(instance, [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] for vif in network_info: [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return self._sync_wrapper(fn, *args, **kwargs) [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self.wait() [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self[:] = self._gt.wait() [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return self._exit_event.wait() [ 631.252837] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] result = hub.switch() [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return self.greenlet.switch() [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] result = function(*args, **kwargs) [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] return func(*args, **kwargs) [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] raise e [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] nwinfo = self.network_api.allocate_for_instance( [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.253358] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] created_port_ids = self._update_ports_for_instance( [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] with excutils.save_and_reraise_exception(): [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] self.force_reraise() [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] raise self.value [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] updated_port = self._update_port( [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] _ensure_no_port_binding_failure(port) [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.253850] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] raise exception.PortBindingFailed(port_id=port['id']) [ 631.254230] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] nova.exception.PortBindingFailed: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. [ 631.254230] env[61936]: ERROR nova.compute.manager [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] [ 631.254230] env[61936]: DEBUG nova.compute.utils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.254230] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.749s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.256682] env[61936]: INFO nova.compute.claims [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.258069] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Build of instance 0257bab6-c040-4f1f-8218-c5bc353777bc was re-scheduled: Binding failed for port a9fad4ea-4497-4063-872f-36e6fe643096, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 631.258495] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 631.258833] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Acquiring lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.259012] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Acquired lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.259206] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.262283] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Releasing lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.262667] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 631.262844] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.263570] env[61936]: DEBUG oslo_concurrency.lockutils [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] Acquired lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.263737] env[61936]: DEBUG nova.network.neutron [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Refreshing network info cache for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 631.264575] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2e2d717-b4a3-4675-90fc-7414046ce90a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.276844] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3af5a5-0941-4221-936a-8877b6fe1528 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.304100] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aad2d5db-9e34-40d3-816c-902eecd0069e could not be found. [ 631.304344] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.304515] env[61936]: INFO nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 631.304812] env[61936]: DEBUG oslo.service.loopingcall [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.305053] env[61936]: DEBUG nova.compute.manager [-] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 631.305205] env[61936]: DEBUG nova.network.neutron [-] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.341216] env[61936]: DEBUG nova.network.neutron [-] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.712730] env[61936]: INFO nova.compute.manager [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: 1944e478-e7ff-4c20-bdfe-95016613c704] Took 1.14 seconds to deallocate network for instance. [ 631.795230] env[61936]: DEBUG nova.network.neutron [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.832233] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.848730] env[61936]: DEBUG nova.network.neutron [-] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.944037] env[61936]: DEBUG nova.network.neutron [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.122770] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.270389] env[61936]: ERROR nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. [ 632.270389] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.270389] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.270389] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.270389] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.270389] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.270389] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.270389] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.270389] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.270389] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 632.270389] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.270389] env[61936]: ERROR nova.compute.manager raise self.value [ 632.270389] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.270389] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.270389] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.270389] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.271162] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.271162] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.271162] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. [ 632.271162] env[61936]: ERROR nova.compute.manager [ 632.271162] env[61936]: Traceback (most recent call last): [ 632.271162] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.271162] env[61936]: listener.cb(fileno) [ 632.271162] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.271162] env[61936]: result = function(*args, **kwargs) [ 632.271162] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.271162] env[61936]: return func(*args, **kwargs) [ 632.271162] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.271162] env[61936]: raise e [ 632.271162] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.271162] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 632.271162] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.271162] env[61936]: created_port_ids = self._update_ports_for_instance( [ 632.271162] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.271162] env[61936]: with excutils.save_and_reraise_exception(): [ 632.271162] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.271162] env[61936]: self.force_reraise() [ 632.271162] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.271162] env[61936]: raise self.value [ 632.271162] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.271162] env[61936]: updated_port = self._update_port( [ 632.271162] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.271162] env[61936]: _ensure_no_port_binding_failure(port) [ 632.271162] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.271162] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.272048] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. [ 632.272048] env[61936]: Removing descriptor: 15 [ 632.272048] env[61936]: ERROR nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Traceback (most recent call last): [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] yield resources [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self.driver.spawn(context, instance, image_meta, [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.272048] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] vm_ref = self.build_virtual_machine(instance, [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] for vif in network_info: [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return self._sync_wrapper(fn, *args, **kwargs) [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self.wait() [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self[:] = self._gt.wait() [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return self._exit_event.wait() [ 632.272430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] result = hub.switch() [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return self.greenlet.switch() [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] result = function(*args, **kwargs) [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return func(*args, **kwargs) [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] raise e [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] nwinfo = self.network_api.allocate_for_instance( [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.272795] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] created_port_ids = self._update_ports_for_instance( [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] with excutils.save_and_reraise_exception(): [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self.force_reraise() [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] raise self.value [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] updated_port = self._update_port( [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] _ensure_no_port_binding_failure(port) [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.273447] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] raise exception.PortBindingFailed(port_id=port['id']) [ 632.273892] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] nova.exception.PortBindingFailed: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. [ 632.273892] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] [ 632.273892] env[61936]: INFO nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Terminating instance [ 632.354175] env[61936]: INFO nova.compute.manager [-] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Took 1.05 seconds to deallocate network for instance. [ 632.356560] env[61936]: DEBUG nova.compute.claims [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.356809] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.447824] env[61936]: DEBUG oslo_concurrency.lockutils [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] Releasing lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.448522] env[61936]: DEBUG nova.compute.manager [req-e5e3a09a-1be3-4c76-9d29-abf548260f46 req-5d33e748-9030-408f-bfea-5a4702a84fcc service nova] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Received event network-vif-deleted-3799ef5b-d8b0-44b9-8ee9-372b8eb07303 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 632.627450] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Releasing lock "refresh_cache-0257bab6-c040-4f1f-8218-c5bc353777bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.627878] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 632.627878] env[61936]: DEBUG nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 632.628250] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.656251] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.695543] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f44599-362d-4613-a8f3-de94d4d8e957 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.704870] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1d30f0-af42-4e41-9afe-dc16e5ee49a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.747110] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4223428-c5ad-4cdd-8595-efa56725d992 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.754369] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce1c739-fa7a-473e-8eab-f11843d827f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.769174] env[61936]: DEBUG nova.compute.provider_tree [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.771211] env[61936]: INFO nova.scheduler.client.report [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Deleted allocations for instance 1944e478-e7ff-4c20-bdfe-95016613c704 [ 632.779413] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Acquiring lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.779629] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Acquired lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.779929] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.881874] env[61936]: DEBUG nova.compute.manager [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Received event network-changed-8bd92dbc-402e-430d-935a-9732f71bcfe9 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 632.881874] env[61936]: DEBUG nova.compute.manager [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Refreshing instance network info cache due to event network-changed-8bd92dbc-402e-430d-935a-9732f71bcfe9. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 632.881874] env[61936]: DEBUG oslo_concurrency.lockutils [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] Acquiring lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.161453] env[61936]: DEBUG nova.network.neutron [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.280742] env[61936]: DEBUG nova.scheduler.client.report [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 633.282076] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8a789aa8-e77b-4212-9bd0-7496c1e45e9f tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "1944e478-e7ff-4c20-bdfe-95016613c704" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.162s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.306041] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.484487] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.646743] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "bdf4c76e-2441-4f76-bf74-2718c57c13e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.646962] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "bdf4c76e-2441-4f76-bf74-2718c57c13e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.663640] env[61936]: INFO nova.compute.manager [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] [instance: 0257bab6-c040-4f1f-8218-c5bc353777bc] Took 1.03 seconds to deallocate network for instance. [ 633.786880] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.788561] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 633.791011] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.584s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.793021] env[61936]: INFO nova.compute.claims [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.799365] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 633.991015] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Releasing lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.991015] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 633.991015] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.991015] env[61936]: DEBUG oslo_concurrency.lockutils [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] Acquired lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.991015] env[61936]: DEBUG nova.network.neutron [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Refreshing network info cache for port 8bd92dbc-402e-430d-935a-9732f71bcfe9 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 633.991630] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73ff76b8-5ba3-410c-a888-80de1702b0e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.011024] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130d52a6-ce39-4c25-944b-d90ab65647a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.036426] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance de5277a7-05e2-49eb-b577-87f900eeef3b could not be found. [ 634.036816] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.037122] env[61936]: INFO nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 634.037495] env[61936]: DEBUG oslo.service.loopingcall [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.037828] env[61936]: DEBUG nova.compute.manager [-] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 634.038343] env[61936]: DEBUG nova.network.neutron [-] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.062750] env[61936]: DEBUG nova.network.neutron [-] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.188468] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "f6c3575f-9351-4957-a200-4e4d0557b55b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.188468] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "f6c3575f-9351-4957-a200-4e4d0557b55b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.292433] env[61936]: DEBUG nova.compute.utils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.294082] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 634.294265] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.329573] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.393438] env[61936]: DEBUG nova.policy [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a0dbd54ae01f474ca6e38f177d0eff9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd04c1c7b46704aa19b6781e081ff8c83', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.566254] env[61936]: DEBUG nova.network.neutron [-] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.605408] env[61936]: DEBUG nova.network.neutron [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.705428] env[61936]: INFO nova.scheduler.client.report [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Deleted allocations for instance 0257bab6-c040-4f1f-8218-c5bc353777bc [ 634.800764] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 634.846545] env[61936]: DEBUG nova.network.neutron [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.069072] env[61936]: INFO nova.compute.manager [-] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Took 1.03 seconds to deallocate network for instance. [ 635.073799] env[61936]: DEBUG nova.compute.claims [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.073973] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.139384] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Successfully created port: 8ce3e12a-8016-46c3-add8-94b23d6abf51 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.150174] env[61936]: DEBUG nova.compute.manager [req-6e8606fc-62c6-463e-b95c-c1db4ecf1107 req-f32ec221-4812-40cd-b86d-68fcb32c55bb service nova] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Received event network-vif-deleted-8bd92dbc-402e-430d-935a-9732f71bcfe9 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 635.208207] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515b4c0d-73f6-4326-a25c-c56da7f633aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.216195] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a352187-282c-4e92-8fff-5b1a9f0a7102 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.219510] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7ec8a2b2-0b27-4142-b5fa-e4546743c00f tempest-FloatingIPsAssociationNegativeTestJSON-1763193901 tempest-FloatingIPsAssociationNegativeTestJSON-1763193901-project-member] Lock "0257bab6-c040-4f1f-8218-c5bc353777bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.461s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.249748] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfedbcaf-5041-4711-a156-1016b5ce6817 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.257852] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b861be3e-5530-4649-bc82-bd45f96341dc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.272127] env[61936]: DEBUG nova.compute.provider_tree [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.351759] env[61936]: DEBUG oslo_concurrency.lockutils [req-dabd3254-30d3-4d15-9b57-cd964f9da6f0 req-33b2261c-2792-460d-95f9-487ac8359521 service nova] Releasing lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.722663] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 635.776022] env[61936]: DEBUG nova.scheduler.client.report [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 635.821326] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 635.849382] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 635.849616] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.852680] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 635.852680] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.852680] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 635.852680] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 635.852680] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 635.853000] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 635.853000] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 635.853000] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 635.853000] env[61936]: DEBUG nova.virt.hardware [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 635.853000] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cce5701-d40d-4c9d-aabc-04f4ce723276 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.861966] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c9283f-bda5-4634-ac4c-17800f20a934 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.250490] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.283191] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.283191] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 636.284331] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.316s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.284851] env[61936]: DEBUG nova.objects.instance [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lazy-loading 'resources' on Instance uuid 7dadef86-7685-40e2-9e68-debc4b4a8a95 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 636.792079] env[61936]: DEBUG nova.compute.utils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.794051] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 636.794234] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.960410] env[61936]: DEBUG nova.policy [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c76388554af4ca7aa63e9e296b4ca52', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9571d89db4940f18176d634ecee9bee', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.227499] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da13a3cd-d0ca-4193-8dcd-0f941987e4ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.234908] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393cbe79-fe52-47b0-8b87-72e9349c0adf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.267185] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4487bea-db89-4fcc-b1fb-bfc9149d76b0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.276350] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69058519-f60e-4c6b-8413-a506588beac3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.297363] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 637.299079] env[61936]: DEBUG nova.compute.provider_tree [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.484045] env[61936]: DEBUG nova.compute.manager [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Received event network-changed-8ce3e12a-8016-46c3-add8-94b23d6abf51 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 637.486597] env[61936]: DEBUG nova.compute.manager [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Refreshing instance network info cache due to event network-changed-8ce3e12a-8016-46c3-add8-94b23d6abf51. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 637.487126] env[61936]: DEBUG oslo_concurrency.lockutils [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] Acquiring lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.487882] env[61936]: DEBUG oslo_concurrency.lockutils [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] Acquired lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.487882] env[61936]: DEBUG nova.network.neutron [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Refreshing network info cache for port 8ce3e12a-8016-46c3-add8-94b23d6abf51 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.801609] env[61936]: DEBUG nova.scheduler.client.report [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 637.878825] env[61936]: ERROR nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. [ 637.878825] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 637.878825] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.878825] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 637.878825] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.878825] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 637.878825] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.878825] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 637.878825] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.878825] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 637.878825] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.878825] env[61936]: ERROR nova.compute.manager raise self.value [ 637.878825] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.878825] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 637.878825] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.878825] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 637.879645] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.879645] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 637.879645] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. [ 637.879645] env[61936]: ERROR nova.compute.manager [ 637.879645] env[61936]: Traceback (most recent call last): [ 637.879645] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 637.879645] env[61936]: listener.cb(fileno) [ 637.879645] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.879645] env[61936]: result = function(*args, **kwargs) [ 637.879645] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.879645] env[61936]: return func(*args, **kwargs) [ 637.879645] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.879645] env[61936]: raise e [ 637.879645] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.879645] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 637.879645] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.879645] env[61936]: created_port_ids = self._update_ports_for_instance( [ 637.879645] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.879645] env[61936]: with excutils.save_and_reraise_exception(): [ 637.879645] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.879645] env[61936]: self.force_reraise() [ 637.879645] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.879645] env[61936]: raise self.value [ 637.879645] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.879645] env[61936]: updated_port = self._update_port( [ 637.879645] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.879645] env[61936]: _ensure_no_port_binding_failure(port) [ 637.879645] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.879645] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 637.880826] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. [ 637.880826] env[61936]: Removing descriptor: 15 [ 637.880826] env[61936]: ERROR nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] Traceback (most recent call last): [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] yield resources [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self.driver.spawn(context, instance, image_meta, [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.880826] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] vm_ref = self.build_virtual_machine(instance, [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] for vif in network_info: [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return self._sync_wrapper(fn, *args, **kwargs) [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self.wait() [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self[:] = self._gt.wait() [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return self._exit_event.wait() [ 637.881606] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] result = hub.switch() [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return self.greenlet.switch() [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] result = function(*args, **kwargs) [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return func(*args, **kwargs) [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] raise e [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] nwinfo = self.network_api.allocate_for_instance( [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.882286] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] created_port_ids = self._update_ports_for_instance( [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] with excutils.save_and_reraise_exception(): [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self.force_reraise() [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] raise self.value [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] updated_port = self._update_port( [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] _ensure_no_port_binding_failure(port) [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.882904] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] raise exception.PortBindingFailed(port_id=port['id']) [ 637.884304] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] nova.exception.PortBindingFailed: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. [ 637.884304] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] [ 637.884304] env[61936]: INFO nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Terminating instance [ 638.025025] env[61936]: DEBUG nova.network.neutron [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.054329] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Successfully created port: a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.194433] env[61936]: DEBUG nova.network.neutron [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.310645] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.026s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.313942] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 638.317738] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.223s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.360022] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 638.360022] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.360022] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 638.360022] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.360442] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 638.360442] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 638.360442] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 638.360442] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 638.360442] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 638.360599] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 638.360599] env[61936]: DEBUG nova.virt.hardware [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 638.360599] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd28b1cf-b323-4586-982b-fe9facf6b396 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.375252] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4fe7707-f249-4ba7-b148-43464f1b6c1d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.395900] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Acquiring lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.397620] env[61936]: INFO nova.scheduler.client.report [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Deleted allocations for instance 7dadef86-7685-40e2-9e68-debc4b4a8a95 [ 638.697030] env[61936]: DEBUG oslo_concurrency.lockutils [req-14f96408-f372-48bd-af3e-873124dc3a15 req-65efee46-cde4-4834-a276-1457cebb6f99 service nova] Releasing lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.697030] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Acquired lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.697030] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.911133] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b685dc6-29c4-4916-87a9-4fb2f642f62d tempest-ServerDiagnosticsV248Test-915812372 tempest-ServerDiagnosticsV248Test-915812372-project-member] Lock "7dadef86-7685-40e2-9e68-debc4b4a8a95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.272s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.259511] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcd08a8-b22c-4cdd-9fa9-4a889268663b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.268407] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7459d8a4-abc4-4fb1-bc7c-e8d9984bc96f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.303449] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.306625] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef6400a-fe48-4efd-bf1f-f46481eb08c4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.316163] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1cdffe-d89d-4e65-8d47-5850f82ffc02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.329721] env[61936]: DEBUG nova.compute.provider_tree [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.551171] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.667537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Acquiring lock "d7599598-492e-4d42-8417-960df12b2cd7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.667802] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Lock "d7599598-492e-4d42-8417-960df12b2cd7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.837090] env[61936]: DEBUG nova.scheduler.client.report [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 640.056515] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Releasing lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.056977] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 640.058637] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.058637] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42d3cf4d-94e7-4036-b3a6-8d9113961ff5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.068720] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2af729d-7b62-44d0-80bb-c21edf9dced0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.096666] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9302959-c96c-47cb-980d-13493aa185f8 could not be found. [ 640.096888] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 640.097078] env[61936]: INFO nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 640.097332] env[61936]: DEBUG oslo.service.loopingcall [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.097554] env[61936]: DEBUG nova.compute.manager [-] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 640.097646] env[61936]: DEBUG nova.network.neutron [-] [instance: c9302959-c96c-47cb-980d-13493aa185f8] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.127507] env[61936]: DEBUG nova.compute.manager [req-4fed0f76-8290-4f73-9edf-f4d3b13b5498 req-0124f305-4f0f-40d2-b680-698f861f0a42 service nova] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Received event network-vif-deleted-8ce3e12a-8016-46c3-add8-94b23d6abf51 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 640.146167] env[61936]: DEBUG nova.network.neutron [-] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.347090] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.347899] env[61936]: ERROR nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Traceback (most recent call last): [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self.driver.spawn(context, instance, image_meta, [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] vm_ref = self.build_virtual_machine(instance, [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.347899] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] for vif in network_info: [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return self._sync_wrapper(fn, *args, **kwargs) [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self.wait() [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self[:] = self._gt.wait() [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return self._exit_event.wait() [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] result = hub.switch() [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.349077] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return self.greenlet.switch() [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] result = function(*args, **kwargs) [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] return func(*args, **kwargs) [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] raise e [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] nwinfo = self.network_api.allocate_for_instance( [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] created_port_ids = self._update_ports_for_instance( [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] with excutils.save_and_reraise_exception(): [ 640.349512] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] self.force_reraise() [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] raise self.value [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] updated_port = self._update_port( [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] _ensure_no_port_binding_failure(port) [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] raise exception.PortBindingFailed(port_id=port['id']) [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] nova.exception.PortBindingFailed: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. [ 640.349884] env[61936]: ERROR nova.compute.manager [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] [ 640.350222] env[61936]: DEBUG nova.compute.utils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.357019] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.060s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.361064] env[61936]: INFO nova.compute.claims [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.362619] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Build of instance e3f6ec50-6d1b-46bd-889d-efe70bc7b98f was re-scheduled: Binding failed for port 732af640-3bab-48bb-a44d-efc86f420c9a, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 640.363292] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 640.363673] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquiring lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.363965] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Acquired lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.365403] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.651442] env[61936]: DEBUG nova.network.neutron [-] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.864349] env[61936]: ERROR nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. [ 640.864349] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.864349] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.864349] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.864349] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.864349] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.864349] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.864349] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.864349] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.864349] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 640.864349] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.864349] env[61936]: ERROR nova.compute.manager raise self.value [ 640.864349] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.864349] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.864349] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.864349] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.864890] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.864890] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.864890] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. [ 640.864890] env[61936]: ERROR nova.compute.manager [ 640.864890] env[61936]: Traceback (most recent call last): [ 640.864890] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.864890] env[61936]: listener.cb(fileno) [ 640.864890] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.864890] env[61936]: result = function(*args, **kwargs) [ 640.864890] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.864890] env[61936]: return func(*args, **kwargs) [ 640.864890] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.864890] env[61936]: raise e [ 640.864890] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.864890] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 640.864890] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.864890] env[61936]: created_port_ids = self._update_ports_for_instance( [ 640.864890] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.864890] env[61936]: with excutils.save_and_reraise_exception(): [ 640.864890] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.864890] env[61936]: self.force_reraise() [ 640.864890] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.864890] env[61936]: raise self.value [ 640.864890] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.864890] env[61936]: updated_port = self._update_port( [ 640.864890] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.864890] env[61936]: _ensure_no_port_binding_failure(port) [ 640.864890] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.864890] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.868413] env[61936]: nova.exception.PortBindingFailed: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. [ 640.868413] env[61936]: Removing descriptor: 19 [ 640.868413] env[61936]: ERROR nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Traceback (most recent call last): [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] yield resources [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self.driver.spawn(context, instance, image_meta, [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.868413] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] vm_ref = self.build_virtual_machine(instance, [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] for vif in network_info: [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return self._sync_wrapper(fn, *args, **kwargs) [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self.wait() [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self[:] = self._gt.wait() [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return self._exit_event.wait() [ 640.868845] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] result = hub.switch() [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return self.greenlet.switch() [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] result = function(*args, **kwargs) [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return func(*args, **kwargs) [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] raise e [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] nwinfo = self.network_api.allocate_for_instance( [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.869271] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] created_port_ids = self._update_ports_for_instance( [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] with excutils.save_and_reraise_exception(): [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self.force_reraise() [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] raise self.value [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] updated_port = self._update_port( [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] _ensure_no_port_binding_failure(port) [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.869770] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] raise exception.PortBindingFailed(port_id=port['id']) [ 640.871620] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] nova.exception.PortBindingFailed: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. [ 640.871620] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] [ 640.871620] env[61936]: INFO nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Terminating instance [ 640.894355] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.085952] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.155999] env[61936]: INFO nova.compute.manager [-] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Took 1.06 seconds to deallocate network for instance. [ 641.162582] env[61936]: DEBUG nova.compute.claims [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.163531] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.372693] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Acquiring lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.372945] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Acquired lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.373105] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.593414] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Releasing lock "refresh_cache-e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.593414] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 641.593414] env[61936]: DEBUG nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.593414] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.615711] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.864253] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022e5907-f6d0-430a-8301-435201443353 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.871730] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc275cc-cf6f-4241-9b05-ad095564c46d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.907422] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0eefb08-64ea-49a3-8c6f-fbc367501e95 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.915820] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943a7da7-95f2-4d48-b6d4-22a87f101436 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.932485] env[61936]: DEBUG nova.compute.provider_tree [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.955853] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.119947] env[61936]: DEBUG nova.network.neutron [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.243800] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.263207] env[61936]: DEBUG nova.compute.manager [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Received event network-changed-a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 642.263416] env[61936]: DEBUG nova.compute.manager [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Refreshing instance network info cache due to event network-changed-a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 642.263599] env[61936]: DEBUG oslo_concurrency.lockutils [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] Acquiring lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.436713] env[61936]: DEBUG nova.scheduler.client.report [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 642.622832] env[61936]: INFO nova.compute.manager [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] [instance: e3f6ec50-6d1b-46bd-889d-efe70bc7b98f] Took 1.03 seconds to deallocate network for instance. [ 642.749804] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Releasing lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.752444] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 642.752444] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 642.752444] env[61936]: DEBUG oslo_concurrency.lockutils [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] Acquired lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.752444] env[61936]: DEBUG nova.network.neutron [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Refreshing network info cache for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 642.752444] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d612e4a-ec3f-4944-9ec9-d3edbab72a0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.765026] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5506defe-b094-482f-a1c9-d0d7bffe9768 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.794415] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69563e1e-301a-4a7b-8fe4-7f1846478f56 could not be found. [ 642.794653] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 642.794833] env[61936]: INFO nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Took 0.04 seconds to destroy the instance on the hypervisor. [ 642.795091] env[61936]: DEBUG oslo.service.loopingcall [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.795331] env[61936]: DEBUG nova.compute.manager [-] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 642.795431] env[61936]: DEBUG nova.network.neutron [-] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.816492] env[61936]: DEBUG nova.network.neutron [-] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.946689] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.947709] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 642.951322] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.074s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.285847] env[61936]: DEBUG nova.network.neutron [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.318599] env[61936]: DEBUG nova.network.neutron [-] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.458550] env[61936]: DEBUG nova.compute.utils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.464546] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 643.464723] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.658375] env[61936]: DEBUG nova.network.neutron [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.663906] env[61936]: INFO nova.scheduler.client.report [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Deleted allocations for instance e3f6ec50-6d1b-46bd-889d-efe70bc7b98f [ 643.821139] env[61936]: INFO nova.compute.manager [-] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Took 1.03 seconds to deallocate network for instance. [ 643.823413] env[61936]: DEBUG nova.compute.claims [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 643.823826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.841711] env[61936]: DEBUG nova.policy [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f32face6793640c8b3d744a16e28cc88', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43e6da536e4f44609298c76c2f4fd477', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.965664] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 644.040782] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46aba36-a0cc-404e-bd4e-26d4f461d714 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.054662] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a09143-a13a-405c-887e-7102e643b4e6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.092855] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc553d9-26dc-49df-99e3-2be8f0ff47ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.102992] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a25639e-0477-4d54-a751-8646d39ee607 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.118120] env[61936]: DEBUG nova.compute.provider_tree [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.173523] env[61936]: DEBUG oslo_concurrency.lockutils [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] Releasing lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.173617] env[61936]: DEBUG nova.compute.manager [req-58fb397c-ba26-40f1-97bd-f208232bab5b req-45093983-5337-4078-8cf0-144fa1353eb7 service nova] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Received event network-vif-deleted-a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 644.175257] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3110140e-d3e4-4c57-9187-1f9c1b8e5b34 tempest-ListServerFiltersTestJSON-1157183276 tempest-ListServerFiltersTestJSON-1157183276-project-member] Lock "e3f6ec50-6d1b-46bd-889d-efe70bc7b98f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.112s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.623524] env[61936]: DEBUG nova.scheduler.client.report [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 644.653539] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Successfully created port: 6ac85da9-c165-4a01-839b-7212f0ac7fa9 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 644.679788] env[61936]: DEBUG nova.compute.manager [None req-c318f2c7-c052-4e2f-a950-1cdb9054f319 tempest-ServersListShow296Test-359310352 tempest-ServersListShow296Test-359310352-project-member] [instance: 62fe885c-6320-4d4a-86f8-4a5afcc6f928] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 644.978267] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 645.012228] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 645.012333] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.012655] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 645.013201] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.013201] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 645.019031] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 645.019031] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 645.019031] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 645.019031] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 645.019031] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 645.019361] env[61936]: DEBUG nova.virt.hardware [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 645.019361] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6f16a3-1838-44d3-ab10-13a5315292c8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.028199] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea945a9-a57c-4b0a-bff4-e0a718c28914 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.132428] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.181s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.133290] env[61936]: ERROR nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Traceback (most recent call last): [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self.driver.spawn(context, instance, image_meta, [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] vm_ref = self.build_virtual_machine(instance, [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.133290] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] for vif in network_info: [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return self._sync_wrapper(fn, *args, **kwargs) [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self.wait() [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self[:] = self._gt.wait() [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return self._exit_event.wait() [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] result = hub.switch() [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.133921] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return self.greenlet.switch() [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] result = function(*args, **kwargs) [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] return func(*args, **kwargs) [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] raise e [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] nwinfo = self.network_api.allocate_for_instance( [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] created_port_ids = self._update_ports_for_instance( [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] with excutils.save_and_reraise_exception(): [ 645.134438] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] self.force_reraise() [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] raise self.value [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] updated_port = self._update_port( [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] _ensure_no_port_binding_failure(port) [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] raise exception.PortBindingFailed(port_id=port['id']) [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] nova.exception.PortBindingFailed: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. [ 645.135888] env[61936]: ERROR nova.compute.manager [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] [ 645.137976] env[61936]: DEBUG nova.compute.utils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 645.137976] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.808s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.137976] env[61936]: INFO nova.compute.claims [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.140364] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Build of instance 39a94d7e-77bb-4083-81c3-4dce3bf55fdd was re-scheduled: Binding failed for port 5fe3de13-f680-43f2-bed5-24cd390263a0, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 645.140842] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 645.141110] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Acquiring lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.141265] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Acquired lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.141424] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.185733] env[61936]: DEBUG nova.compute.manager [None req-c318f2c7-c052-4e2f-a950-1cdb9054f319 tempest-ServersListShow296Test-359310352 tempest-ServersListShow296Test-359310352-project-member] [instance: 62fe885c-6320-4d4a-86f8-4a5afcc6f928] Instance disappeared before build. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2445}} [ 645.689830] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.715192] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c318f2c7-c052-4e2f-a950-1cdb9054f319 tempest-ServersListShow296Test-359310352 tempest-ServersListShow296Test-359310352-project-member] Lock "62fe885c-6320-4d4a-86f8-4a5afcc6f928" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.608s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.862186] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.220949] env[61936]: DEBUG nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 646.366672] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Releasing lock "refresh_cache-39a94d7e-77bb-4083-81c3-4dce3bf55fdd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.368301] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 646.368301] env[61936]: DEBUG nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 646.368301] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.412942] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.486642] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Acquiring lock "201d06f9-d277-4d8e-84ee-84e1053c1137" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.486880] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Lock "201d06f9-d277-4d8e-84ee-84e1053c1137" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.637019] env[61936]: ERROR nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. [ 646.637019] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.637019] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.637019] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.637019] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.637019] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.637019] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.637019] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.637019] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.637019] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 646.637019] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.637019] env[61936]: ERROR nova.compute.manager raise self.value [ 646.637019] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.637019] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.637019] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.637019] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.637449] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.637449] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.637449] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. [ 646.637449] env[61936]: ERROR nova.compute.manager [ 646.637449] env[61936]: Traceback (most recent call last): [ 646.637449] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.637449] env[61936]: listener.cb(fileno) [ 646.637449] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.637449] env[61936]: result = function(*args, **kwargs) [ 646.637449] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.637449] env[61936]: return func(*args, **kwargs) [ 646.637449] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 646.637449] env[61936]: raise e [ 646.637449] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.637449] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 646.637449] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.637449] env[61936]: created_port_ids = self._update_ports_for_instance( [ 646.637449] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.637449] env[61936]: with excutils.save_and_reraise_exception(): [ 646.637449] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.637449] env[61936]: self.force_reraise() [ 646.637449] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.637449] env[61936]: raise self.value [ 646.637449] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.637449] env[61936]: updated_port = self._update_port( [ 646.637449] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.637449] env[61936]: _ensure_no_port_binding_failure(port) [ 646.637449] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.637449] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.638073] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. [ 646.638073] env[61936]: Removing descriptor: 15 [ 646.638073] env[61936]: ERROR nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Traceback (most recent call last): [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] yield resources [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self.driver.spawn(context, instance, image_meta, [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.638073] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] vm_ref = self.build_virtual_machine(instance, [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] for vif in network_info: [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return self._sync_wrapper(fn, *args, **kwargs) [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self.wait() [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self[:] = self._gt.wait() [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return self._exit_event.wait() [ 646.638439] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] result = hub.switch() [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return self.greenlet.switch() [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] result = function(*args, **kwargs) [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return func(*args, **kwargs) [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] raise e [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] nwinfo = self.network_api.allocate_for_instance( [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.638725] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] created_port_ids = self._update_ports_for_instance( [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] with excutils.save_and_reraise_exception(): [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self.force_reraise() [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] raise self.value [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] updated_port = self._update_port( [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] _ensure_no_port_binding_failure(port) [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.639078] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] raise exception.PortBindingFailed(port_id=port['id']) [ 646.641144] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] nova.exception.PortBindingFailed: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. [ 646.641144] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] [ 646.641144] env[61936]: INFO nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Terminating instance [ 646.643045] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5695d2-853b-42dc-843b-5c45c769b8f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.651596] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af8a0da-dc02-4f9d-aa57-bce39590f55f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.682407] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44cc0fb-d2a5-4ee0-b438-a02fb5c437f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.690323] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb9b032-4589-4402-be69-936c35d4232d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.705645] env[61936]: DEBUG nova.compute.provider_tree [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.741155] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.917387] env[61936]: DEBUG nova.network.neutron [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.993867] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Acquiring lock "03ab71e9-64b7-4534-a693-473f3903c511" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.994117] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Lock "03ab71e9-64b7-4534-a693-473f3903c511" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.141127] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Acquiring lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.141318] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Acquired lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.141499] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.206591] env[61936]: DEBUG nova.scheduler.client.report [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 647.284658] env[61936]: DEBUG nova.compute.manager [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Received event network-changed-6ac85da9-c165-4a01-839b-7212f0ac7fa9 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 647.284933] env[61936]: DEBUG nova.compute.manager [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Refreshing instance network info cache due to event network-changed-6ac85da9-c165-4a01-839b-7212f0ac7fa9. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 647.285221] env[61936]: DEBUG oslo_concurrency.lockutils [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] Acquiring lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.423149] env[61936]: INFO nova.compute.manager [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] [instance: 39a94d7e-77bb-4083-81c3-4dce3bf55fdd] Took 1.05 seconds to deallocate network for instance. [ 647.654146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Acquiring lock "8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.654146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Lock "8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.656240] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Acquiring lock "0b14583c-dd0f-429d-a386-2acda353465a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.656433] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Lock "0b14583c-dd0f-429d-a386-2acda353465a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.678606] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.712729] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.713021] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 647.715864] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.359s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.842666] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.218436] env[61936]: DEBUG nova.compute.utils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.220492] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 648.221137] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.313170] env[61936]: DEBUG nova.policy [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd6bc18d304f4c19a115235cab0e288a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e4408cc87c942b082b305e4b4f3ab23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.350094] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Releasing lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.350559] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 648.350767] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.351160] env[61936]: DEBUG oslo_concurrency.lockutils [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] Acquired lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.351367] env[61936]: DEBUG nova.network.neutron [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Refreshing network info cache for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 648.352639] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79aed713-973c-445b-bd0a-06d7b6c675f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.370853] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0a0a95-7dbf-46cd-a1c2-464ff24cfd17 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.399958] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ac0103e-8424-4518-8acb-343267deb3ab could not be found. [ 648.400411] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.400411] env[61936]: INFO nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Took 0.05 seconds to destroy the instance on the hypervisor. [ 648.400628] env[61936]: DEBUG oslo.service.loopingcall [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.403612] env[61936]: DEBUG nova.compute.manager [-] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 648.403742] env[61936]: DEBUG nova.network.neutron [-] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.445374] env[61936]: DEBUG nova.network.neutron [-] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.466417] env[61936]: INFO nova.scheduler.client.report [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Deleted allocations for instance 39a94d7e-77bb-4083-81c3-4dce3bf55fdd [ 648.727581] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 648.791279] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3e5c84-859f-476f-af8f-68ab322cef92 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.801478] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756cf9bf-0e7c-470c-9e05-cb8f4eaab7d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.838734] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da441fba-d822-440e-bc73-e53acffa8341 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.846762] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76884ef3-6afd-4035-850b-984b32d9aa46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.864262] env[61936]: DEBUG nova.compute.provider_tree [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.894453] env[61936]: DEBUG nova.network.neutron [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.948104] env[61936]: DEBUG nova.network.neutron [-] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.977046] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0349186b-4b5f-451d-b066-225bc7df4480 tempest-VolumesAssistedSnapshotsTest-1815446704 tempest-VolumesAssistedSnapshotsTest-1815446704-project-member] Lock "39a94d7e-77bb-4083-81c3-4dce3bf55fdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.625s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.033323] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Successfully created port: f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.137243] env[61936]: DEBUG nova.network.neutron [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.366459] env[61936]: DEBUG nova.scheduler.client.report [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 649.434203] env[61936]: DEBUG nova.compute.manager [req-8257ffc5-37ba-4a88-a6b3-2fe486db049f req-777d00d8-96c0-4e74-81dd-3dfad1cc8aae service nova] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Received event network-vif-deleted-6ac85da9-c165-4a01-839b-7212f0ac7fa9 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 649.451435] env[61936]: INFO nova.compute.manager [-] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Took 1.05 seconds to deallocate network for instance. [ 649.455385] env[61936]: DEBUG nova.compute.claims [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.455645] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.479817] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 649.640367] env[61936]: DEBUG oslo_concurrency.lockutils [req-58a51007-a09c-4af6-997b-494f7cd05571 req-5231beee-7424-4ace-b1f9-189396fa6921 service nova] Releasing lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.745011] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 649.777481] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 649.777736] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.777889] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 649.778076] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.778233] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 649.778485] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 649.778700] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 649.778858] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 649.779973] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 649.780239] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 649.780451] env[61936]: DEBUG nova.virt.hardware [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 649.781409] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f6c594-762b-4581-a172-892540e96f26 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.790007] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3226d78a-49b2-44b6-b140-f0761c1b0ae5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.875045] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.159s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.876144] env[61936]: ERROR nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Traceback (most recent call last): [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self.driver.spawn(context, instance, image_meta, [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] vm_ref = self.build_virtual_machine(instance, [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.876144] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] for vif in network_info: [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return self._sync_wrapper(fn, *args, **kwargs) [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self.wait() [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self[:] = self._gt.wait() [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return self._exit_event.wait() [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] result = hub.switch() [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.876454] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return self.greenlet.switch() [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] result = function(*args, **kwargs) [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] return func(*args, **kwargs) [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] raise e [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] nwinfo = self.network_api.allocate_for_instance( [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] created_port_ids = self._update_ports_for_instance( [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] with excutils.save_and_reraise_exception(): [ 649.876741] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] self.force_reraise() [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] raise self.value [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] updated_port = self._update_port( [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] _ensure_no_port_binding_failure(port) [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] raise exception.PortBindingFailed(port_id=port['id']) [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] nova.exception.PortBindingFailed: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. [ 649.877049] env[61936]: ERROR nova.compute.manager [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] [ 649.878405] env[61936]: DEBUG nova.compute.utils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 649.881254] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.549s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.883035] env[61936]: INFO nova.compute.claims [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.889558] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Build of instance aad2d5db-9e34-40d3-816c-902eecd0069e was re-scheduled: Binding failed for port 3799ef5b-d8b0-44b9-8ee9-372b8eb07303, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 649.889799] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 649.890531] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.890531] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquired lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.890531] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.010438] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.421061] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.488735] env[61936]: ERROR nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. [ 650.488735] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 650.488735] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.488735] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 650.488735] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.488735] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 650.488735] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.488735] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 650.488735] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.488735] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 650.488735] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.488735] env[61936]: ERROR nova.compute.manager raise self.value [ 650.488735] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.488735] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 650.488735] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.488735] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 650.489094] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.489094] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 650.489094] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. [ 650.489094] env[61936]: ERROR nova.compute.manager [ 650.489094] env[61936]: Traceback (most recent call last): [ 650.489094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 650.489094] env[61936]: listener.cb(fileno) [ 650.489094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.489094] env[61936]: result = function(*args, **kwargs) [ 650.489094] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.489094] env[61936]: return func(*args, **kwargs) [ 650.489094] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 650.489094] env[61936]: raise e [ 650.489094] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.489094] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 650.489094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.489094] env[61936]: created_port_ids = self._update_ports_for_instance( [ 650.489094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.489094] env[61936]: with excutils.save_and_reraise_exception(): [ 650.489094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.489094] env[61936]: self.force_reraise() [ 650.489094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.489094] env[61936]: raise self.value [ 650.489094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.489094] env[61936]: updated_port = self._update_port( [ 650.489094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.489094] env[61936]: _ensure_no_port_binding_failure(port) [ 650.489094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.489094] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 650.489814] env[61936]: nova.exception.PortBindingFailed: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. [ 650.489814] env[61936]: Removing descriptor: 15 [ 650.489814] env[61936]: ERROR nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Traceback (most recent call last): [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] yield resources [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self.driver.spawn(context, instance, image_meta, [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.489814] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] vm_ref = self.build_virtual_machine(instance, [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] for vif in network_info: [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return self._sync_wrapper(fn, *args, **kwargs) [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self.wait() [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self[:] = self._gt.wait() [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return self._exit_event.wait() [ 650.490081] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] result = hub.switch() [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return self.greenlet.switch() [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] result = function(*args, **kwargs) [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return func(*args, **kwargs) [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] raise e [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] nwinfo = self.network_api.allocate_for_instance( [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.490350] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] created_port_ids = self._update_ports_for_instance( [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] with excutils.save_and_reraise_exception(): [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self.force_reraise() [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] raise self.value [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] updated_port = self._update_port( [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] _ensure_no_port_binding_failure(port) [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.490653] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] raise exception.PortBindingFailed(port_id=port['id']) [ 650.490906] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] nova.exception.PortBindingFailed: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. [ 650.490906] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] [ 650.490906] env[61936]: INFO nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Terminating instance [ 650.576281] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.996045] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Acquiring lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.996320] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Acquired lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.996320] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.078683] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Releasing lock "refresh_cache-aad2d5db-9e34-40d3-816c-902eecd0069e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.078886] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 651.079357] env[61936]: DEBUG nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 651.079357] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.096339] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.302119] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c116b3d9-bc3e-4294-8650-ac4e35ba2345 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.310083] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d931c5-1728-4b78-982c-43598aabe850 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.343563] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012c530d-a85b-4235-ae51-748a30db57fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.351678] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086243ae-8713-42ee-9369-85ccfa538f7a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.365797] env[61936]: DEBUG nova.compute.provider_tree [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.491646] env[61936]: DEBUG nova.compute.manager [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Received event network-changed-f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 651.491838] env[61936]: DEBUG nova.compute.manager [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Refreshing instance network info cache due to event network-changed-f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 651.492088] env[61936]: DEBUG oslo_concurrency.lockutils [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] Acquiring lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.517140] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.600600] env[61936]: DEBUG nova.network.neutron [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.627903] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.658640] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.658918] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.871645] env[61936]: DEBUG nova.scheduler.client.report [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 652.108655] env[61936]: INFO nova.compute.manager [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: aad2d5db-9e34-40d3-816c-902eecd0069e] Took 1.03 seconds to deallocate network for instance. [ 652.133668] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Releasing lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.134334] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 652.134632] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.135056] env[61936]: DEBUG oslo_concurrency.lockutils [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] Acquired lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.139027] env[61936]: DEBUG nova.network.neutron [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Refreshing network info cache for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 652.139027] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-977e4002-6ccb-4e9f-a8a2-f7bb2ba229a0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.149887] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba3c10d-6fcc-4226-a51b-9bc2054f4499 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.178790] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab8313a4-b427-456b-ab81-19a0b0f95d71 could not be found. [ 652.179024] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.179204] env[61936]: INFO nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Took 0.04 seconds to destroy the instance on the hypervisor. [ 652.179446] env[61936]: DEBUG oslo.service.loopingcall [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.179663] env[61936]: DEBUG nova.compute.manager [-] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 652.179750] env[61936]: DEBUG nova.network.neutron [-] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.213801] env[61936]: DEBUG nova.network.neutron [-] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.378975] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.378975] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 652.386028] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.310s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.666361] env[61936]: DEBUG nova.network.neutron [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.719206] env[61936]: DEBUG nova.network.neutron [-] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.888190] env[61936]: DEBUG nova.compute.utils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.888190] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 652.888190] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 653.012664] env[61936]: DEBUG nova.network.neutron [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.061314] env[61936]: DEBUG nova.policy [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d5fceff495040bc85f2e0097893036a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3263691079b947d18103f3be0aff5564', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 653.146711] env[61936]: INFO nova.scheduler.client.report [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Deleted allocations for instance aad2d5db-9e34-40d3-816c-902eecd0069e [ 653.223120] env[61936]: INFO nova.compute.manager [-] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Took 1.04 seconds to deallocate network for instance. [ 653.226110] env[61936]: DEBUG nova.compute.claims [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.226268] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.288487] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a66717-4f8b-474d-8fb7-9ae6b8d2ee6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.296601] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23e96d4-afeb-42fc-a4d6-08d6211043ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.334792] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581ba674-1750-4f14-9105-ce2443a35822 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.342386] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e35a22d-eec1-450b-b85f-8ebaab831cb0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.356428] env[61936]: DEBUG nova.compute.provider_tree [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.394517] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 653.515298] env[61936]: DEBUG oslo_concurrency.lockutils [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] Releasing lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.515654] env[61936]: DEBUG nova.compute.manager [req-c0554056-0928-4d89-a667-dcc1a0cafefa req-17b2f120-5479-4827-867e-d1dd60bc3561 service nova] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Received event network-vif-deleted-f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 653.539751] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Successfully created port: af96abca-f5b6-44bf-9e5b-af06a9b3b157 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.664677] env[61936]: DEBUG oslo_concurrency.lockutils [None req-409dd769-eb75-4901-8a4d-7ff413b2d7d4 tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "aad2d5db-9e34-40d3-816c-902eecd0069e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.841s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.861903] env[61936]: DEBUG nova.scheduler.client.report [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 654.170357] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 654.368357] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.369121] env[61936]: ERROR nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Traceback (most recent call last): [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self.driver.spawn(context, instance, image_meta, [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] vm_ref = self.build_virtual_machine(instance, [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.369121] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] for vif in network_info: [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return self._sync_wrapper(fn, *args, **kwargs) [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self.wait() [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self[:] = self._gt.wait() [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return self._exit_event.wait() [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] result = hub.switch() [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.369430] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return self.greenlet.switch() [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] result = function(*args, **kwargs) [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] return func(*args, **kwargs) [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] raise e [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] nwinfo = self.network_api.allocate_for_instance( [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] created_port_ids = self._update_ports_for_instance( [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] with excutils.save_and_reraise_exception(): [ 654.369730] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] self.force_reraise() [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] raise self.value [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] updated_port = self._update_port( [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] _ensure_no_port_binding_failure(port) [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] raise exception.PortBindingFailed(port_id=port['id']) [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] nova.exception.PortBindingFailed: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. [ 654.370034] env[61936]: ERROR nova.compute.manager [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] [ 654.371021] env[61936]: DEBUG nova.compute.utils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.371551] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.122s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.373042] env[61936]: INFO nova.compute.claims [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.378777] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Build of instance de5277a7-05e2-49eb-b577-87f900eeef3b was re-scheduled: Binding failed for port 8bd92dbc-402e-430d-935a-9732f71bcfe9, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 654.382245] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 654.385407] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Acquiring lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.385407] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Acquired lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.385407] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.404176] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 654.437352] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 654.437781] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.437781] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 654.438259] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.438259] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 654.438417] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 654.438921] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 654.438921] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 654.438921] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 654.439111] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 654.439429] env[61936]: DEBUG nova.virt.hardware [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 654.440368] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e576f8-3f17-4cb2-9fc2-3e352e2cce1d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.450724] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3747841-c791-4471-954c-14ef3a60aa6c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.670097] env[61936]: DEBUG nova.compute.manager [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Received event network-changed-af96abca-f5b6-44bf-9e5b-af06a9b3b157 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 654.670301] env[61936]: DEBUG nova.compute.manager [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Refreshing instance network info cache due to event network-changed-af96abca-f5b6-44bf-9e5b-af06a9b3b157. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 654.670525] env[61936]: DEBUG oslo_concurrency.lockutils [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] Acquiring lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.670678] env[61936]: DEBUG oslo_concurrency.lockutils [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] Acquired lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.670837] env[61936]: DEBUG nova.network.neutron [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Refreshing network info cache for port af96abca-f5b6-44bf-9e5b-af06a9b3b157 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 654.693132] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.870475] env[61936]: ERROR nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. [ 654.870475] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.870475] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.870475] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.870475] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.870475] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.870475] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.870475] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.870475] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.870475] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 654.870475] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.870475] env[61936]: ERROR nova.compute.manager raise self.value [ 654.870475] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.870475] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.870475] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.870475] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.870881] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.870881] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.870881] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. [ 654.870881] env[61936]: ERROR nova.compute.manager [ 654.870881] env[61936]: Traceback (most recent call last): [ 654.870881] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.870881] env[61936]: listener.cb(fileno) [ 654.870881] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.870881] env[61936]: result = function(*args, **kwargs) [ 654.870881] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.870881] env[61936]: return func(*args, **kwargs) [ 654.870881] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 654.870881] env[61936]: raise e [ 654.870881] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.870881] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 654.870881] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.870881] env[61936]: created_port_ids = self._update_ports_for_instance( [ 654.870881] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.870881] env[61936]: with excutils.save_and_reraise_exception(): [ 654.870881] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.870881] env[61936]: self.force_reraise() [ 654.870881] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.870881] env[61936]: raise self.value [ 654.870881] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.870881] env[61936]: updated_port = self._update_port( [ 654.870881] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.870881] env[61936]: _ensure_no_port_binding_failure(port) [ 654.870881] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.870881] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.871526] env[61936]: nova.exception.PortBindingFailed: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. [ 654.871526] env[61936]: Removing descriptor: 15 [ 654.871526] env[61936]: ERROR nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Traceback (most recent call last): [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] yield resources [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self.driver.spawn(context, instance, image_meta, [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.871526] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] vm_ref = self.build_virtual_machine(instance, [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] for vif in network_info: [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return self._sync_wrapper(fn, *args, **kwargs) [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self.wait() [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self[:] = self._gt.wait() [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return self._exit_event.wait() [ 654.871787] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] result = hub.switch() [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return self.greenlet.switch() [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] result = function(*args, **kwargs) [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return func(*args, **kwargs) [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] raise e [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] nwinfo = self.network_api.allocate_for_instance( [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.872079] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] created_port_ids = self._update_ports_for_instance( [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] with excutils.save_and_reraise_exception(): [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self.force_reraise() [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] raise self.value [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] updated_port = self._update_port( [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] _ensure_no_port_binding_failure(port) [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.872353] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] raise exception.PortBindingFailed(port_id=port['id']) [ 654.872669] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] nova.exception.PortBindingFailed: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. [ 654.872669] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] [ 654.872669] env[61936]: INFO nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Terminating instance [ 654.908962] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.055979] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.200722] env[61936]: DEBUG nova.network.neutron [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.311115] env[61936]: DEBUG nova.network.neutron [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.375676] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.558868] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Releasing lock "refresh_cache-de5277a7-05e2-49eb-b577-87f900eeef3b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.559101] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 655.559674] env[61936]: DEBUG nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 655.559674] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.574586] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.816189] env[61936]: DEBUG oslo_concurrency.lockutils [req-e16feb01-8b0d-4a85-b6ca-b9eb2acbf551 req-f36eec74-7227-4597-b46f-c9d1d0b3fd35 service nova] Releasing lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.818048] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquired lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.818338] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.822702] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.822702] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.843032] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b8a006-6583-403f-ae50-aab19c936629 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.851300] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfef8f0-e1b9-4b6f-8af6-654f3b96ea2b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.885961] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0debb14-deae-4a79-b66b-698d061d5b6c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.893768] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf9ebef-2902-48f2-9d69-82b8a102816c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.906945] env[61936]: DEBUG nova.compute.provider_tree [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.083149] env[61936]: DEBUG nova.network.neutron [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.362031] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.410280] env[61936]: DEBUG nova.scheduler.client.report [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.497026] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.591037] env[61936]: INFO nova.compute.manager [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] [instance: de5277a7-05e2-49eb-b577-87f900eeef3b] Took 1.03 seconds to deallocate network for instance. [ 656.735201] env[61936]: DEBUG nova.compute.manager [req-23f6488c-6cd7-44bc-ac8e-107a41ae4208 req-8f077a5a-5b1d-4a93-bb45-c63918fc4684 service nova] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Received event network-vif-deleted-af96abca-f5b6-44bf-9e5b-af06a9b3b157 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 656.918159] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.918159] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 656.919974] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.758s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.003076] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Releasing lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.003511] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 657.003700] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 657.003973] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6eea85cc-5247-47db-ae4b-e31a0749e9b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.012627] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a9ce92-497b-4ccc-869b-90778e5446b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.034245] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fac84980-b2ba-4de8-afed-228a28a0f299 could not be found. [ 657.034470] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.034689] env[61936]: INFO nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Took 0.03 seconds to destroy the instance on the hypervisor. [ 657.034944] env[61936]: DEBUG oslo.service.loopingcall [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 657.035164] env[61936]: DEBUG nova.compute.manager [-] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 657.035259] env[61936]: DEBUG nova.network.neutron [-] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.050738] env[61936]: DEBUG nova.network.neutron [-] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.426905] env[61936]: DEBUG nova.compute.utils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.431839] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 657.431839] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 657.508315] env[61936]: DEBUG nova.policy [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a27e6c20c5b149d49e46c27089bd33d2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24243173e2c645d19b91b59c9ce678bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.553765] env[61936]: DEBUG nova.network.neutron [-] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.630074] env[61936]: INFO nova.scheduler.client.report [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Deleted allocations for instance de5277a7-05e2-49eb-b577-87f900eeef3b [ 657.826699] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832e7fb4-15bb-4e9a-b5f2-3da3ff33e9bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.833502] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c74f63b-340a-41a3-86cc-40de054f9282 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.862949] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685411eb-51b2-4adf-8986-213ec58519bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.870031] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f43f61-65b3-4d03-a926-b7a08dd11370 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.882949] env[61936]: DEBUG nova.compute.provider_tree [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.932962] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 658.052642] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Successfully created port: a78a7314-6023-4650-a7ac-583240c3f102 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.058091] env[61936]: INFO nova.compute.manager [-] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Took 1.02 seconds to deallocate network for instance. [ 658.063074] env[61936]: DEBUG nova.compute.claims [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 658.063221] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.142604] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b5744d0-215e-441b-8f20-4ce262a81010 tempest-ServersWithSpecificFlavorTestJSON-634341115 tempest-ServersWithSpecificFlavorTestJSON-634341115-project-member] Lock "de5277a7-05e2-49eb-b577-87f900eeef3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.044s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.387112] env[61936]: DEBUG nova.scheduler.client.report [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 658.645534] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 658.892093] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.973s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.892729] env[61936]: ERROR nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] Traceback (most recent call last): [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self.driver.spawn(context, instance, image_meta, [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] vm_ref = self.build_virtual_machine(instance, [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.892729] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] for vif in network_info: [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return self._sync_wrapper(fn, *args, **kwargs) [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self.wait() [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self[:] = self._gt.wait() [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return self._exit_event.wait() [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] result = hub.switch() [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.893066] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return self.greenlet.switch() [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] result = function(*args, **kwargs) [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] return func(*args, **kwargs) [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] raise e [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] nwinfo = self.network_api.allocate_for_instance( [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] created_port_ids = self._update_ports_for_instance( [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] with excutils.save_and_reraise_exception(): [ 658.893373] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] self.force_reraise() [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] raise self.value [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] updated_port = self._update_port( [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] _ensure_no_port_binding_failure(port) [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] raise exception.PortBindingFailed(port_id=port['id']) [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] nova.exception.PortBindingFailed: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. [ 658.893675] env[61936]: ERROR nova.compute.manager [instance: c9302959-c96c-47cb-980d-13493aa185f8] [ 658.894384] env[61936]: DEBUG nova.compute.utils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.895786] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.072s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.898695] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Build of instance c9302959-c96c-47cb-980d-13493aa185f8 was re-scheduled: Binding failed for port 8ce3e12a-8016-46c3-add8-94b23d6abf51, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 658.899234] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 658.899712] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Acquiring lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.899712] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Acquired lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.899872] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.941711] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 658.971256] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 658.971522] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.971679] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 658.971857] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.972576] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 658.972576] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 658.972576] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 658.972706] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 658.973033] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 658.973033] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 658.973166] env[61936]: DEBUG nova.virt.hardware [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 658.974403] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572f1b3d-c3b5-4ad6-987e-031d58343960 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.984117] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6048d321-7ae0-4932-b64b-60d805b04fdf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.168312] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.453426] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.625666] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.667733] env[61936]: DEBUG nova.compute.manager [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Received event network-changed-a78a7314-6023-4650-a7ac-583240c3f102 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 659.667988] env[61936]: DEBUG nova.compute.manager [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Refreshing instance network info cache due to event network-changed-a78a7314-6023-4650-a7ac-583240c3f102. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 659.668147] env[61936]: DEBUG oslo_concurrency.lockutils [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] Acquiring lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.668287] env[61936]: DEBUG oslo_concurrency.lockutils [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] Acquired lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.668443] env[61936]: DEBUG nova.network.neutron [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Refreshing network info cache for port a78a7314-6023-4650-a7ac-583240c3f102 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 659.849918] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da13b9c-894b-4a6e-9ed6-002935477dda {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.860856] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0d8c43-23f1-43a4-80ea-5fc08ad1351c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.905294] env[61936]: ERROR nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. [ 659.905294] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 659.905294] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 659.905294] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 659.905294] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.905294] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 659.905294] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.905294] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 659.905294] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.905294] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 659.905294] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.905294] env[61936]: ERROR nova.compute.manager raise self.value [ 659.905294] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.905294] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 659.905294] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.905294] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 659.905754] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.905754] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 659.905754] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. [ 659.905754] env[61936]: ERROR nova.compute.manager [ 659.905754] env[61936]: Traceback (most recent call last): [ 659.905754] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 659.905754] env[61936]: listener.cb(fileno) [ 659.905754] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.905754] env[61936]: result = function(*args, **kwargs) [ 659.905754] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.905754] env[61936]: return func(*args, **kwargs) [ 659.905754] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 659.905754] env[61936]: raise e [ 659.905754] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 659.905754] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 659.905754] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.905754] env[61936]: created_port_ids = self._update_ports_for_instance( [ 659.905754] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.905754] env[61936]: with excutils.save_and_reraise_exception(): [ 659.905754] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.905754] env[61936]: self.force_reraise() [ 659.905754] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.905754] env[61936]: raise self.value [ 659.905754] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.905754] env[61936]: updated_port = self._update_port( [ 659.905754] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.905754] env[61936]: _ensure_no_port_binding_failure(port) [ 659.905754] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.905754] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 659.906445] env[61936]: nova.exception.PortBindingFailed: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. [ 659.906445] env[61936]: Removing descriptor: 15 [ 659.906445] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4a92b4-cae4-4f59-aa9e-b850557a45ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.909877] env[61936]: ERROR nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Traceback (most recent call last): [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] yield resources [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self.driver.spawn(context, instance, image_meta, [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] vm_ref = self.build_virtual_machine(instance, [ 659.909877] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] for vif in network_info: [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return self._sync_wrapper(fn, *args, **kwargs) [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self.wait() [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self[:] = self._gt.wait() [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return self._exit_event.wait() [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 659.910213] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] result = hub.switch() [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return self.greenlet.switch() [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] result = function(*args, **kwargs) [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return func(*args, **kwargs) [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] raise e [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] nwinfo = self.network_api.allocate_for_instance( [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] created_port_ids = self._update_ports_for_instance( [ 659.910493] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] with excutils.save_and_reraise_exception(): [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self.force_reraise() [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] raise self.value [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] updated_port = self._update_port( [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] _ensure_no_port_binding_failure(port) [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] raise exception.PortBindingFailed(port_id=port['id']) [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] nova.exception.PortBindingFailed: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. [ 659.910770] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] [ 659.911071] env[61936]: INFO nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Terminating instance [ 659.917052] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e687f70-ea04-4a29-937d-7e6d4b670ad7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.933426] env[61936]: DEBUG nova.compute.provider_tree [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.127775] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Releasing lock "refresh_cache-c9302959-c96c-47cb-980d-13493aa185f8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.128012] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 660.128191] env[61936]: DEBUG nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 660.128357] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.147292] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.193688] env[61936]: DEBUG nova.network.neutron [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.341520] env[61936]: DEBUG nova.network.neutron [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.414419] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Acquiring lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.436687] env[61936]: DEBUG nova.scheduler.client.report [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 660.654032] env[61936]: DEBUG nova.network.neutron [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.844427] env[61936]: DEBUG oslo_concurrency.lockutils [req-c0f4e7eb-f618-4144-a469-11f0293978e2 req-afbc6c9f-9ed9-4c25-8734-951a5c4c5eb3 service nova] Releasing lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.844890] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Acquired lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.845043] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.942981] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.943670] env[61936]: ERROR nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Traceback (most recent call last): [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self.driver.spawn(context, instance, image_meta, [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] vm_ref = self.build_virtual_machine(instance, [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.943670] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] for vif in network_info: [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return self._sync_wrapper(fn, *args, **kwargs) [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self.wait() [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self[:] = self._gt.wait() [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return self._exit_event.wait() [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] result = hub.switch() [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.944629] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return self.greenlet.switch() [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] result = function(*args, **kwargs) [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] return func(*args, **kwargs) [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] raise e [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] nwinfo = self.network_api.allocate_for_instance( [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] created_port_ids = self._update_ports_for_instance( [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] with excutils.save_and_reraise_exception(): [ 660.945575] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] self.force_reraise() [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] raise self.value [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] updated_port = self._update_port( [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] _ensure_no_port_binding_failure(port) [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] raise exception.PortBindingFailed(port_id=port['id']) [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] nova.exception.PortBindingFailed: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. [ 660.946397] env[61936]: ERROR nova.compute.manager [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] [ 660.946714] env[61936]: DEBUG nova.compute.utils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.946714] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.205s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.947289] env[61936]: INFO nova.compute.claims [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.950884] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Build of instance 69563e1e-301a-4a7b-8fe4-7f1846478f56 was re-scheduled: Binding failed for port a47430bc-0247-4eb2-a5ec-2bf2a3d5a43c, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 660.950884] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 660.950884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Acquiring lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.950884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Acquired lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.951174] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.159342] env[61936]: INFO nova.compute.manager [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] [instance: c9302959-c96c-47cb-980d-13493aa185f8] Took 1.03 seconds to deallocate network for instance. [ 661.374723] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.491264] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.601214] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.689245] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.330680] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Releasing lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.331111] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 662.331304] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.333908] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Releasing lock "refresh_cache-69563e1e-301a-4a7b-8fe4-7f1846478f56" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.334250] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 662.334361] env[61936]: DEBUG nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 662.334413] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.336543] env[61936]: DEBUG nova.compute.manager [req-c47960ef-5447-40ab-92d3-2e1a9f6356f7 req-20dd999f-30e1-4d2a-9860-fcc4200e73f3 service nova] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Received event network-vif-deleted-a78a7314-6023-4650-a7ac-583240c3f102 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 662.339130] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58397730-9edf-48f9-85eb-f240eae33a6a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.348266] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b39b228-f31a-4479-9506-34d97d72ae58 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.361265] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.371224] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ecf38466-be6f-432a-bdf4-f2403e53c024 could not be found. [ 662.371671] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.371779] env[61936]: INFO nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Took 0.04 seconds to destroy the instance on the hypervisor. [ 662.371978] env[61936]: DEBUG oslo.service.loopingcall [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.372277] env[61936]: DEBUG nova.compute.manager [-] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 662.372369] env[61936]: DEBUG nova.network.neutron [-] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.386762] env[61936]: INFO nova.scheduler.client.report [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Deleted allocations for instance c9302959-c96c-47cb-980d-13493aa185f8 [ 662.401019] env[61936]: DEBUG nova.network.neutron [-] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.776704] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1c6791-8bd6-47b8-a3e3-0d37f33cdffd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.786332] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9135ff49-8e20-49d6-bea4-25f9e0b5ba18 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.818931] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468ee8ef-9dc1-450d-89f0-d830f52ea18f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.827848] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7d36e9-e96a-4abb-9fc0-a21c6cd54c87 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.842477] env[61936]: DEBUG nova.compute.provider_tree [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.867571] env[61936]: DEBUG nova.network.neutron [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.894745] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a5b5b163-352d-455e-bb60-8266a6a3d86b tempest-TenantUsagesTestJSON-770683776 tempest-TenantUsagesTestJSON-770683776-project-member] Lock "c9302959-c96c-47cb-980d-13493aa185f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.057s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.902732] env[61936]: DEBUG nova.network.neutron [-] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.348254] env[61936]: DEBUG nova.scheduler.client.report [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 663.370714] env[61936]: INFO nova.compute.manager [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] [instance: 69563e1e-301a-4a7b-8fe4-7f1846478f56] Took 1.04 seconds to deallocate network for instance. [ 663.401062] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 663.406505] env[61936]: INFO nova.compute.manager [-] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Took 1.03 seconds to deallocate network for instance. [ 663.409569] env[61936]: DEBUG nova.compute.claims [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.409793] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.852902] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.907s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.854319] env[61936]: DEBUG nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 663.857819] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.402s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.924826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.363026] env[61936]: DEBUG nova.compute.utils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 664.367615] env[61936]: DEBUG nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Not allocating networking since 'none' was specified. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 664.418918] env[61936]: INFO nova.scheduler.client.report [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Deleted allocations for instance 69563e1e-301a-4a7b-8fe4-7f1846478f56 [ 664.750078] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c399101-9ec0-46c9-b46a-2130c5b45541 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.757923] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b039f241-606d-474d-ae81-55c4260d8645 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.788699] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d8bd40e-fed6-4f7f-89ec-1a96b01b4dbb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.796322] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06efcf1-60d3-4d23-818a-98f044c977ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.810063] env[61936]: DEBUG nova.compute.provider_tree [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.869749] env[61936]: DEBUG nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 664.928714] env[61936]: DEBUG oslo_concurrency.lockutils [None req-794d92a8-81e9-4b8d-8737-e495ffc2c2eb tempest-ServerRescueTestJSON-75128735 tempest-ServerRescueTestJSON-75128735-project-member] Lock "69563e1e-301a-4a7b-8fe4-7f1846478f56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.973s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.313052] env[61936]: DEBUG nova.scheduler.client.report [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 665.433784] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 665.819779] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.820458] env[61936]: ERROR nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Traceback (most recent call last): [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self.driver.spawn(context, instance, image_meta, [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] vm_ref = self.build_virtual_machine(instance, [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.820458] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] for vif in network_info: [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return self._sync_wrapper(fn, *args, **kwargs) [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self.wait() [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self[:] = self._gt.wait() [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return self._exit_event.wait() [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] result = hub.switch() [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.820719] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return self.greenlet.switch() [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] result = function(*args, **kwargs) [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] return func(*args, **kwargs) [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] raise e [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] nwinfo = self.network_api.allocate_for_instance( [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] created_port_ids = self._update_ports_for_instance( [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] with excutils.save_and_reraise_exception(): [ 665.821234] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] self.force_reraise() [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] raise self.value [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] updated_port = self._update_port( [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] _ensure_no_port_binding_failure(port) [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] raise exception.PortBindingFailed(port_id=port['id']) [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] nova.exception.PortBindingFailed: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. [ 665.821594] env[61936]: ERROR nova.compute.manager [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] [ 665.821845] env[61936]: DEBUG nova.compute.utils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.826766] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Build of instance 9ac0103e-8424-4518-8acb-343267deb3ab was re-scheduled: Binding failed for port 6ac85da9-c165-4a01-839b-7212f0ac7fa9, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 665.826766] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 665.826766] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Acquiring lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.826766] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Acquired lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.826766] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.827481] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.817s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.828938] env[61936]: INFO nova.compute.claims [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.878580] env[61936]: DEBUG nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 665.911672] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 665.912485] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.912485] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 665.912485] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.912485] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 665.912631] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 665.912701] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 665.913055] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 665.913055] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 665.913219] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 665.913437] env[61936]: DEBUG nova.virt.hardware [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 665.914291] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e7c7aa-a438-425b-b1b2-c7ce3bc29036 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.922570] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2ab176-f2b3-478f-803b-6b0ac3103449 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.939901] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 665.946594] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Creating folder: Project (45b30a5dd2624a8ab325b75474efc361). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 665.947898] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6f221d3-ef42-4c30-87bc-f0581ba2514c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.958180] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Created folder: Project (45b30a5dd2624a8ab325b75474efc361) in parent group-v269874. [ 665.958973] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Creating folder: Instances. Parent ref: group-v269882. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 665.958973] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83c2b7f4-026a-4326-a505-e3b9a697855e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.970105] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Created folder: Instances in parent group-v269882. [ 665.970105] env[61936]: DEBUG oslo.service.loopingcall [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.970105] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.970105] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 665.970105] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-575de776-f727-4f60-b111-85bfa4ba89dd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.985826] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 665.985826] env[61936]: value = "task-1252719" [ 665.985826] env[61936]: _type = "Task" [ 665.985826] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.993412] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252719, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.381781] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.496736] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252719, 'name': CreateVM_Task, 'duration_secs': 0.232042} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.496736] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 666.497268] env[61936]: DEBUG oslo_vmware.service [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2270baf-c6c8-47af-91dc-bdb186c4ad92 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.506177] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.506177] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.506177] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 666.506177] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ade26061-c13a-4e7b-af4b-bdae92302016 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.508614] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 666.508614] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52123be6-0493-a2b9-a03e-c9d20b4c6ab1" [ 666.508614] env[61936]: _type = "Task" [ 666.508614] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.515883] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52123be6-0493-a2b9-a03e-c9d20b4c6ab1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.518498] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.024028] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Releasing lock "refresh_cache-9ac0103e-8424-4518-8acb-343267deb3ab" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.027474] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 667.027474] env[61936]: DEBUG nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 667.027474] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.027474] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.027474] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 667.027810] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.027810] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.027810] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 667.027893] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2d6e405-1420-40c5-a517-9686b1a2eb8d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.050355] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 667.050542] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 667.051693] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.053409] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db50e7f-9c2d-4fe6-81e1-7fa462ee36d0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.064346] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e6d10dd-5f4d-47de-8d58-7df412048026 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.069620] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 667.069620] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526a5ea2-23f9-da0d-af7f-aff4b1ed7ef6" [ 667.069620] env[61936]: _type = "Task" [ 667.069620] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.077953] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526a5ea2-23f9-da0d-af7f-aff4b1ed7ef6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.265888] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1fdb5c-efe7-4ef7-b856-a0b793132f42 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.273834] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18b5db4-fd51-45e4-ae4e-d0279a351e17 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.304829] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31a642d-6828-440a-a327-a333c8fdfd1e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.315044] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e43227-0f6d-4fdc-9dab-78b0a0bb8533 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.327280] env[61936]: DEBUG nova.compute.provider_tree [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.559483] env[61936]: DEBUG nova.network.neutron [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.580038] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Preparing fetch location {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 667.580311] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Creating directory with path [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 667.580544] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6ce044f-53cd-4e2f-b2b6-1abf45443033 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.600819] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Created directory with path [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 667.601095] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Fetch image to [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 667.601272] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Downloading image file data 43acc3d3-5e18-42a0-9168-cb6831c6bbfb to [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk on the data store datastore2 {{(pid=61936) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 667.602088] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19ecec1-5243-4a3d-96fd-3872be214c43 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.610368] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4a7fe2-c6ed-4654-9bd6-c4d5f39cc2a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.620520] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519c7aad-dbc8-4401-bf72-e20108db53af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.652394] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785e33c8-650f-4319-9ff4-beeb1c65395d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.658995] env[61936]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-3116c7bf-bcd9-4a3a-b724-93f840ec2540 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.747739] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Downloading image file data 43acc3d3-5e18-42a0-9168-cb6831c6bbfb to the data store datastore2 {{(pid=61936) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 667.827791] env[61936]: DEBUG oslo_vmware.rw_handles [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61936) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 667.832972] env[61936]: DEBUG nova.scheduler.client.report [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 667.893258] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.066s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.893786] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 667.896229] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.670s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.062209] env[61936]: INFO nova.compute.manager [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] [instance: 9ac0103e-8424-4518-8acb-343267deb3ab] Took 1.04 seconds to deallocate network for instance. [ 668.404154] env[61936]: DEBUG nova.compute.utils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.409231] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 668.409231] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 668.493815] env[61936]: DEBUG nova.policy [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d5fceff495040bc85f2e0097893036a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3263691079b947d18103f3be0aff5564', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.499017] env[61936]: DEBUG oslo_vmware.rw_handles [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Completed reading data from the image iterator. {{(pid=61936) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 668.499017] env[61936]: DEBUG oslo_vmware.rw_handles [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 668.638748] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Downloaded image file data 43acc3d3-5e18-42a0-9168-cb6831c6bbfb to vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk on the data store datastore2 {{(pid=61936) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 668.640809] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Caching image {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 668.644304] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Copying Virtual Disk [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk to [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 668.645752] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-641f634e-774c-4322-9ce1-f443f3855843 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.656563] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 668.656563] env[61936]: value = "task-1252720" [ 668.656563] env[61936]: _type = "Task" [ 668.656563] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.664170] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252720, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.828295] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45755aa3-ebdc-41e4-a8fa-e802d3da45b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.836212] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09cff9bc-77dc-4250-8109-76bb531c953a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.872249] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbce53d-ac0f-4664-a1b8-129e5aac7414 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.879986] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abeac62-777a-47df-bf40-f516eef2ff53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.893417] env[61936]: DEBUG nova.compute.provider_tree [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.913633] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 668.995651] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Successfully created port: 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.110569] env[61936]: INFO nova.scheduler.client.report [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Deleted allocations for instance 9ac0103e-8424-4518-8acb-343267deb3ab [ 669.171125] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252720, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.396359] env[61936]: DEBUG nova.scheduler.client.report [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 669.622274] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6936d903-b5e0-4135-a1e5-586b977f9bfe tempest-ServersTestJSON-139160545 tempest-ServersTestJSON-139160545-project-member] Lock "9ac0103e-8424-4518-8acb-343267deb3ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.889s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.636204] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Acquiring lock "4755f16d-254b-47e9-8b50-05df95690445" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.636911] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Lock "4755f16d-254b-47e9-8b50-05df95690445" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.670706] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252720, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635607} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.671228] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Copied Virtual Disk [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk to [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 669.671412] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleting the datastore file [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/tmp-sparse.vmdk {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 669.671667] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93d1e011-0304-40d4-a800-94c91fda448d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.678245] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 669.678245] env[61936]: value = "task-1252721" [ 669.678245] env[61936]: _type = "Task" [ 669.678245] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.688452] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.902475] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.905016] env[61936]: ERROR nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Traceback (most recent call last): [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self.driver.spawn(context, instance, image_meta, [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] vm_ref = self.build_virtual_machine(instance, [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.905016] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] for vif in network_info: [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return self._sync_wrapper(fn, *args, **kwargs) [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self.wait() [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self[:] = self._gt.wait() [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return self._exit_event.wait() [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] result = hub.switch() [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.905345] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return self.greenlet.switch() [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] result = function(*args, **kwargs) [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] return func(*args, **kwargs) [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] raise e [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] nwinfo = self.network_api.allocate_for_instance( [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] created_port_ids = self._update_ports_for_instance( [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] with excutils.save_and_reraise_exception(): [ 669.905676] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] self.force_reraise() [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] raise self.value [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] updated_port = self._update_port( [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] _ensure_no_port_binding_failure(port) [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] raise exception.PortBindingFailed(port_id=port['id']) [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] nova.exception.PortBindingFailed: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. [ 669.905981] env[61936]: ERROR nova.compute.manager [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] [ 669.906319] env[61936]: DEBUG nova.compute.utils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.906794] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Build of instance ab8313a4-b427-456b-ab81-19a0b0f95d71 was re-scheduled: Binding failed for port f9d5a2e8-4fe0-43e0-aee6-c4ede0c8626e, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 669.907348] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 669.907746] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Acquiring lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.908016] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Acquired lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.910020] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.910020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.216s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.912158] env[61936]: INFO nova.compute.claims [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 669.927669] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 669.965842] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 669.966456] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.966456] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 669.966456] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.966456] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 669.966627] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 669.966785] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 669.966932] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 669.967100] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 669.967255] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 669.967490] env[61936]: DEBUG nova.virt.hardware [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 669.969342] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191d9e2b-3c85-444f-bce3-4af11442d042 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.979981] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03cc4efa-5469-471b-a15a-91467f3022de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.125333] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 670.192698] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.046306} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.192698] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 670.192698] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Moving file from [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e/43acc3d3-5e18-42a0-9168-cb6831c6bbfb to [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb. {{(pid=61936) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 670.192951] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-68957c7d-6947-479a-9632-55d21b554914 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.207211] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 670.207211] env[61936]: value = "task-1252722" [ 670.207211] env[61936]: _type = "Task" [ 670.207211] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.217368] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252722, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.451870] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.625518] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.650659] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.693443] env[61936]: ERROR nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. [ 670.693443] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.693443] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 670.693443] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.693443] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.693443] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.693443] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.693443] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.693443] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.693443] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 670.693443] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.693443] env[61936]: ERROR nova.compute.manager raise self.value [ 670.693443] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.693443] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.693443] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.693443] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.693956] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.693956] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.693956] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. [ 670.693956] env[61936]: ERROR nova.compute.manager [ 670.693956] env[61936]: Traceback (most recent call last): [ 670.693956] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.693956] env[61936]: listener.cb(fileno) [ 670.693956] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.693956] env[61936]: result = function(*args, **kwargs) [ 670.693956] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.693956] env[61936]: return func(*args, **kwargs) [ 670.693956] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 670.693956] env[61936]: raise e [ 670.693956] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 670.693956] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 670.693956] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.693956] env[61936]: created_port_ids = self._update_ports_for_instance( [ 670.693956] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.693956] env[61936]: with excutils.save_and_reraise_exception(): [ 670.693956] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.693956] env[61936]: self.force_reraise() [ 670.693956] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.693956] env[61936]: raise self.value [ 670.693956] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.693956] env[61936]: updated_port = self._update_port( [ 670.693956] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.693956] env[61936]: _ensure_no_port_binding_failure(port) [ 670.693956] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.693956] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.694602] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. [ 670.694602] env[61936]: Removing descriptor: 19 [ 670.694771] env[61936]: ERROR nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Traceback (most recent call last): [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] yield resources [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self.driver.spawn(context, instance, image_meta, [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] vm_ref = self.build_virtual_machine(instance, [ 670.694771] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] for vif in network_info: [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return self._sync_wrapper(fn, *args, **kwargs) [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self.wait() [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self[:] = self._gt.wait() [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return self._exit_event.wait() [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.695061] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] result = hub.switch() [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return self.greenlet.switch() [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] result = function(*args, **kwargs) [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return func(*args, **kwargs) [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] raise e [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] nwinfo = self.network_api.allocate_for_instance( [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] created_port_ids = self._update_ports_for_instance( [ 670.695473] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] with excutils.save_and_reraise_exception(): [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self.force_reraise() [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] raise self.value [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] updated_port = self._update_port( [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] _ensure_no_port_binding_failure(port) [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] raise exception.PortBindingFailed(port_id=port['id']) [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] nova.exception.PortBindingFailed: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. [ 670.695818] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] [ 670.696141] env[61936]: INFO nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Terminating instance [ 670.718494] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252722, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026456} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.719024] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] File moved {{(pid=61936) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 670.719456] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Cleaning up location [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 670.719456] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleting the datastore file [datastore2] vmware_temp/c3d0aed4-a38c-462d-a2ec-824e3e87016e {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 670.719729] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfa3b1d8-d97d-4993-9ed0-0dc776791a02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.726898] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 670.726898] env[61936]: value = "task-1252723" [ 670.726898] env[61936]: _type = "Task" [ 670.726898] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.736092] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.750648] env[61936]: DEBUG nova.compute.manager [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Received event network-changed-38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 670.750648] env[61936]: DEBUG nova.compute.manager [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Refreshing instance network info cache due to event network-changed-38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 670.750648] env[61936]: DEBUG oslo_concurrency.lockutils [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] Acquiring lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.750648] env[61936]: DEBUG oslo_concurrency.lockutils [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] Acquired lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.750961] env[61936]: DEBUG nova.network.neutron [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Refreshing network info cache for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 670.801051] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "2fdb605e-72d5-4eef-bab2-0601e730243a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.801507] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "2fdb605e-72d5-4eef-bab2-0601e730243a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.130198] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Releasing lock "refresh_cache-ab8313a4-b427-456b-ab81-19a0b0f95d71" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.130198] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 671.130198] env[61936]: DEBUG nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 671.130198] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.169334] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.203443] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.238907] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024741} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.239193] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 671.241259] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1123d51a-33c4-4fe0-ae83-075230d4f69a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.245248] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 671.245248] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d079d5-d9b6-7893-5a4d-81fae9316e08" [ 671.245248] env[61936]: _type = "Task" [ 671.245248] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.258473] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d079d5-d9b6-7893-5a4d-81fae9316e08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.296704] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2a3b12-0faa-40a0-a164-44f932bfcec1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.302145] env[61936]: DEBUG nova.network.neutron [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.307893] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2748b378-0094-45bb-9326-e02b24229a0a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.344898] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cda6eb-9652-4689-b616-cff6940c596c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.352764] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.353110] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.357707] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a384c0e-14a7-4c32-8d0f-a1e19082d09e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.383108] env[61936]: DEBUG nova.compute.provider_tree [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.478131] env[61936]: DEBUG nova.network.neutron [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.671590] env[61936]: DEBUG nova.network.neutron [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.757769] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d079d5-d9b6-7893-5a4d-81fae9316e08, 'name': SearchDatastore_Task, 'duration_secs': 0.010669} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.758353] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.758770] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 671.759184] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2dafbbbf-f0bc-43a6-b960-1869b729b8c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.765372] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 671.765372] env[61936]: value = "task-1252724" [ 671.765372] env[61936]: _type = "Task" [ 671.765372] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.773823] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252724, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.889039] env[61936]: DEBUG nova.scheduler.client.report [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 671.980850] env[61936]: DEBUG oslo_concurrency.lockutils [req-48f2f497-f15d-4079-8160-9d0e608d4927 req-afc675e1-81e1-46fa-a37e-920ee561ff18 service nova] Releasing lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.981268] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquired lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.981453] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.175202] env[61936]: INFO nova.compute.manager [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] [instance: ab8313a4-b427-456b-ab81-19a0b0f95d71] Took 1.05 seconds to deallocate network for instance. [ 672.277626] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252724, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.397170] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.397170] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 672.399200] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.336s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.509149] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.656164] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.775515] env[61936]: DEBUG nova.compute.manager [req-d2b198f9-321b-4ac0-854b-788e3e4c518d req-6a6b5877-e955-4218-911f-de251c80a8d6 service nova] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Received event network-vif-deleted-38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 672.779434] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252724, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.710476} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.779686] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 672.779895] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 672.780149] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9edd21f5-0a33-4382-b898-b05e0cc1d87d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.787321] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 672.787321] env[61936]: value = "task-1252725" [ 672.787321] env[61936]: _type = "Task" [ 672.787321] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.795696] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252725, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.908553] env[61936]: DEBUG nova.compute.utils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 672.910064] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 672.910235] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 673.012041] env[61936]: DEBUG nova.policy [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06199ff338204a119f8402413843e22d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fa23de03fc343fa94b8c58e2f928f34', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.074104] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.074404] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.162236] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Releasing lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.162655] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 673.162914] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.166215] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa1e85ed-e49f-47fe-97ed-63c9ce388df2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.177137] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47bfa14-7126-4f09-9b7a-7f60a3171958 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.204496] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2 could not be found. [ 673.204727] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 673.204946] env[61936]: INFO nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 673.205215] env[61936]: DEBUG oslo.service.loopingcall [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.205453] env[61936]: DEBUG nova.compute.manager [-] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 673.205546] env[61936]: DEBUG nova.network.neutron [-] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.222979] env[61936]: INFO nova.scheduler.client.report [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Deleted allocations for instance ab8313a4-b427-456b-ab81-19a0b0f95d71 [ 673.302697] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252725, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065068} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.303830] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 673.304482] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95340d12-2176-4e3b-8c0b-9ee372847218 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.308195] env[61936]: DEBUG nova.network.neutron [-] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.328211] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 673.332915] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dc840f6-0d0e-4e1f-b514-845c870fd2a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.351756] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 673.351756] env[61936]: value = "task-1252726" [ 673.351756] env[61936]: _type = "Task" [ 673.351756] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.367391] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.393048] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2538f3b-8114-4b1d-9f95-106dec578bec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.400343] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f9a63e-495d-41e8-9899-e40c7cac949b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.435493] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 673.439167] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d411f4-c639-493e-b215-6ae347446077 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.446369] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0524525c-99b2-4e3a-82cf-6df78f4d4f0d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.462185] env[61936]: DEBUG nova.compute.provider_tree [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.583327] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.583327] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 673.583447] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Rebuilding the list of instances to heal {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10330}} [ 673.718157] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Successfully created port: 3e890892-15a4-4ec1-969e-ca60c92588cc {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.738899] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6a9607e3-641c-44a6-99e4-44af2d286d21 tempest-FloatingIPsAssociationTestJSON-87365 tempest-FloatingIPsAssociationTestJSON-87365-project-member] Lock "ab8313a4-b427-456b-ab81-19a0b0f95d71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.287s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.810718] env[61936]: DEBUG nova.network.neutron [-] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.863573] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252726, 'name': ReconfigVM_Task, 'duration_secs': 0.321903} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.864053] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 673.864773] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a2e7fd8-5eb6-42cb-86db-3b4f8c973eb0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.873888] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 673.873888] env[61936]: value = "task-1252727" [ 673.873888] env[61936]: _type = "Task" [ 673.873888] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.881665] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252727, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.965358] env[61936]: DEBUG nova.scheduler.client.report [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 674.088386] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 674.088669] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 674.090019] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 674.090019] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 674.090019] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Didn't find any instances for network info cache update. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10412}} [ 674.090019] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.090019] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.090019] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.090308] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.090308] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.090609] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.091041] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 674.091336] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.244714] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 674.313516] env[61936]: INFO nova.compute.manager [-] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Took 1.11 seconds to deallocate network for instance. [ 674.316064] env[61936]: DEBUG nova.compute.claims [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 674.316064] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.383727] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252727, 'name': Rename_Task, 'duration_secs': 0.122226} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.384247] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 674.384392] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a86066e1-4223-4e8e-9cb1-831abe1969d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.390340] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 674.390340] env[61936]: value = "task-1252728" [ 674.390340] env[61936]: _type = "Task" [ 674.390340] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.399678] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.450196] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 674.474749] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.075s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.479925] env[61936]: ERROR nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Traceback (most recent call last): [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self.driver.spawn(context, instance, image_meta, [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] vm_ref = self.build_virtual_machine(instance, [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.479925] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] for vif in network_info: [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return self._sync_wrapper(fn, *args, **kwargs) [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self.wait() [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self[:] = self._gt.wait() [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return self._exit_event.wait() [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] result = hub.switch() [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 674.480364] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return self.greenlet.switch() [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] result = function(*args, **kwargs) [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] return func(*args, **kwargs) [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] raise e [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] nwinfo = self.network_api.allocate_for_instance( [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] created_port_ids = self._update_ports_for_instance( [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] with excutils.save_and_reraise_exception(): [ 674.480713] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] self.force_reraise() [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] raise self.value [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] updated_port = self._update_port( [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] _ensure_no_port_binding_failure(port) [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] raise exception.PortBindingFailed(port_id=port['id']) [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] nova.exception.PortBindingFailed: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. [ 674.481074] env[61936]: ERROR nova.compute.manager [instance: fac84980-b2ba-4de8-afed-228a28a0f299] [ 674.481382] env[61936]: DEBUG nova.compute.utils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.481382] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.309s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.481382] env[61936]: INFO nova.compute.claims [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.485043] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Build of instance fac84980-b2ba-4de8-afed-228a28a0f299 was re-scheduled: Binding failed for port af96abca-f5b6-44bf-9e5b-af06a9b3b157, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 674.485043] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 674.485043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.485043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquired lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.485325] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.512798] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 674.512922] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.512983] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 674.513178] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.513321] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 674.513466] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 674.513670] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 674.513821] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 674.513982] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 674.514229] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 674.514345] env[61936]: DEBUG nova.virt.hardware [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 674.515528] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf8ee3a-2644-49e1-9d70-0a7c4f95710e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.528152] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9f3035-61b1-40c0-841c-4ea0e5db7115 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.594699] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.768861] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.901261] env[61936]: DEBUG oslo_vmware.api [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252728, 'name': PowerOnVM_Task, 'duration_secs': 0.452933} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.901576] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 674.901724] env[61936]: INFO nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Took 9.02 seconds to spawn the instance on the hypervisor. [ 674.901892] env[61936]: DEBUG nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 674.902675] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8561336f-82c6-4e80-9b2e-c3bdb5476bd7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.031960] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.292683] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.424188] env[61936]: INFO nova.compute.manager [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Took 28.70 seconds to build instance. [ 675.750040] env[61936]: DEBUG nova.compute.manager [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Received event network-changed-3e890892-15a4-4ec1-969e-ca60c92588cc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 675.750040] env[61936]: DEBUG nova.compute.manager [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Refreshing instance network info cache due to event network-changed-3e890892-15a4-4ec1-969e-ca60c92588cc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 675.750040] env[61936]: DEBUG oslo_concurrency.lockutils [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] Acquiring lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.750040] env[61936]: DEBUG oslo_concurrency.lockutils [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] Acquired lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.750527] env[61936]: DEBUG nova.network.neutron [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Refreshing network info cache for port 3e890892-15a4-4ec1-969e-ca60c92588cc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 675.799305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Releasing lock "refresh_cache-fac84980-b2ba-4de8-afed-228a28a0f299" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.799305] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 675.799305] env[61936]: DEBUG nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 675.799305] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.839720] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.916210] env[61936]: ERROR nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. [ 675.916210] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.916210] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 675.916210] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.916210] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.916210] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.916210] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.916210] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.916210] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.916210] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 675.916210] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.916210] env[61936]: ERROR nova.compute.manager raise self.value [ 675.916210] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.916210] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.916210] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.916210] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.916850] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.916850] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.916850] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. [ 675.916850] env[61936]: ERROR nova.compute.manager [ 675.916850] env[61936]: Traceback (most recent call last): [ 675.916850] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.916850] env[61936]: listener.cb(fileno) [ 675.916850] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.916850] env[61936]: result = function(*args, **kwargs) [ 675.916850] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.916850] env[61936]: return func(*args, **kwargs) [ 675.916850] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 675.916850] env[61936]: raise e [ 675.916850] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 675.916850] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 675.916850] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.916850] env[61936]: created_port_ids = self._update_ports_for_instance( [ 675.916850] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.916850] env[61936]: with excutils.save_and_reraise_exception(): [ 675.916850] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.916850] env[61936]: self.force_reraise() [ 675.916850] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.916850] env[61936]: raise self.value [ 675.916850] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.916850] env[61936]: updated_port = self._update_port( [ 675.916850] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.916850] env[61936]: _ensure_no_port_binding_failure(port) [ 675.916850] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.916850] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.917667] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. [ 675.917667] env[61936]: Removing descriptor: 19 [ 675.917667] env[61936]: ERROR nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Traceback (most recent call last): [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] yield resources [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self.driver.spawn(context, instance, image_meta, [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.917667] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] vm_ref = self.build_virtual_machine(instance, [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] for vif in network_info: [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return self._sync_wrapper(fn, *args, **kwargs) [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self.wait() [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self[:] = self._gt.wait() [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return self._exit_event.wait() [ 675.917949] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] result = hub.switch() [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return self.greenlet.switch() [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] result = function(*args, **kwargs) [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return func(*args, **kwargs) [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] raise e [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] nwinfo = self.network_api.allocate_for_instance( [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.918302] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] created_port_ids = self._update_ports_for_instance( [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] with excutils.save_and_reraise_exception(): [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self.force_reraise() [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] raise self.value [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] updated_port = self._update_port( [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] _ensure_no_port_binding_failure(port) [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.918644] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] raise exception.PortBindingFailed(port_id=port['id']) [ 675.918976] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] nova.exception.PortBindingFailed: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. [ 675.918976] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] [ 675.918976] env[61936]: INFO nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Terminating instance [ 675.928096] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c2a2b4a7-dbbe-46d3-9b3d-bdcbdba3d47e tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.374s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.953166] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afbf840-c85a-4697-b8aa-e528b7abea4b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.964752] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7e3f76-e9a4-4d3b-ba24-bbbf4aa311f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.998882] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad04715a-30f0-4c26-bf11-3dfdc54d63dd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.007992] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f36c879-3ade-4333-9aec-448b53653fc2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.021767] env[61936]: DEBUG nova.compute.provider_tree [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.280655] env[61936]: DEBUG nova.network.neutron [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.344624] env[61936]: DEBUG nova.network.neutron [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.371291] env[61936]: DEBUG nova.network.neutron [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.426822] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.429370] env[61936]: DEBUG nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 676.464560] env[61936]: INFO nova.compute.manager [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Rebuilding instance [ 676.525086] env[61936]: DEBUG nova.scheduler.client.report [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 676.565422] env[61936]: DEBUG nova.compute.manager [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 676.566788] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e7c2a8-a96c-4123-8d63-bc796003f703 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.848609] env[61936]: INFO nova.compute.manager [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: fac84980-b2ba-4de8-afed-228a28a0f299] Took 1.05 seconds to deallocate network for instance. [ 676.877190] env[61936]: DEBUG oslo_concurrency.lockutils [req-bebeef0b-db92-486c-a183-32f72a4dd4b1 req-d03fb4f2-9d88-45fd-a0c2-d1b44629e2dc service nova] Releasing lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.877190] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquired lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.877190] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.958220] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.034922] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.035491] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 677.038876] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.629s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.406125] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.487827] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.540567] env[61936]: DEBUG nova.compute.utils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.544256] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 677.544256] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 677.590251] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 677.590884] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-038be7f5-8327-4798-8343-ca9ad2bde136 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.599255] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 677.599255] env[61936]: value = "task-1252729" [ 677.599255] env[61936]: _type = "Task" [ 677.599255] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.610253] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.651598] env[61936]: DEBUG nova.policy [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '379d69552ef54c30bdc0df5e1738eab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f8d287d9bcf741b39f89ff06d1016b7b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 677.783849] env[61936]: DEBUG nova.compute.manager [req-6ef7bc5a-704b-4efc-a31e-b2e7c86c030b req-f54f0fdf-76a4-4044-9d33-2bb2718a0c25 service nova] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Received event network-vif-deleted-3e890892-15a4-4ec1-969e-ca60c92588cc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 677.847112] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "1251e7f5-684c-428f-9d63-60b77084d09e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.848214] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "1251e7f5-684c-428f-9d63-60b77084d09e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.886873] env[61936]: INFO nova.scheduler.client.report [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Deleted allocations for instance fac84980-b2ba-4de8-afed-228a28a0f299 [ 677.990952] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Releasing lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.991244] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 677.991351] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 677.992695] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-036b9fac-5637-4f11-8914-0186a91b2cd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.997349] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4eff7e-5293-4bfd-992d-e20d61605473 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.007878] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bb8a3d-4561-489d-84f0-0d2b9179a166 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.033229] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542591cd-d5b4-4e1f-956a-94e2b1f4118b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.036766] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 64412fc6-956f-43a6-bd5c-55745360e480 could not be found. [ 678.036964] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.037211] env[61936]: INFO nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Took 0.05 seconds to destroy the instance on the hypervisor. [ 678.037392] env[61936]: DEBUG oslo.service.loopingcall [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.037593] env[61936]: DEBUG nova.compute.manager [-] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 678.037685] env[61936]: DEBUG nova.network.neutron [-] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.068301] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 678.072646] env[61936]: DEBUG nova.network.neutron [-] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.077116] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db00a89a-e47d-4015-bed5-93e6295359c9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.084487] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16181fe2-7506-4275-8ca6-dccce0a911f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.099088] env[61936]: DEBUG nova.compute.provider_tree [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.111735] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252729, 'name': PowerOffVM_Task, 'duration_secs': 0.122091} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.112634] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 678.112875] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 678.113643] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a737bb5-27e6-4fbc-85f3-f6507fab62b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.120828] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 678.121035] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a864ca4-6bcd-4d8f-a04f-0687d1cb98e8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.131264] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Successfully created port: 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.148182] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 678.148742] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 678.149060] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleting the datastore file [datastore2] 994ff54f-a26b-40e0-a2c7-502f3598d9bf {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 678.149426] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-355fb996-df01-4f07-9931-bb20c45e3872 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.157048] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 678.157048] env[61936]: value = "task-1252731" [ 678.157048] env[61936]: _type = "Task" [ 678.157048] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.164498] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252731, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.399369] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aadecabc-21aa-4036-b6d0-7728aced0007 tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "fac84980-b2ba-4de8-afed-228a28a0f299" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.177s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.578545] env[61936]: DEBUG nova.network.neutron [-] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.602332] env[61936]: DEBUG nova.scheduler.client.report [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 678.666729] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252731, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111959} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.667038] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 678.667190] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 678.667338] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.900832] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 679.078718] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 679.083527] env[61936]: INFO nova.compute.manager [-] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Took 1.04 seconds to deallocate network for instance. [ 679.084221] env[61936]: DEBUG nova.compute.claims [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 679.084284] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.107644] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.108981] env[61936]: ERROR nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Traceback (most recent call last): [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self.driver.spawn(context, instance, image_meta, [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] vm_ref = self.build_virtual_machine(instance, [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.108981] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] for vif in network_info: [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return self._sync_wrapper(fn, *args, **kwargs) [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self.wait() [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self[:] = self._gt.wait() [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return self._exit_event.wait() [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] result = hub.switch() [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.109301] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return self.greenlet.switch() [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] result = function(*args, **kwargs) [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] return func(*args, **kwargs) [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] raise e [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] nwinfo = self.network_api.allocate_for_instance( [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] created_port_ids = self._update_ports_for_instance( [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] with excutils.save_and_reraise_exception(): [ 679.109602] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] self.force_reraise() [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] raise self.value [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] updated_port = self._update_port( [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] _ensure_no_port_binding_failure(port) [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] raise exception.PortBindingFailed(port_id=port['id']) [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] nova.exception.PortBindingFailed: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. [ 679.109992] env[61936]: ERROR nova.compute.manager [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] [ 679.110257] env[61936]: DEBUG nova.compute.utils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.113492] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.189s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.118179] env[61936]: INFO nova.compute.claims [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.122771] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Build of instance ecf38466-be6f-432a-bdf4-f2403e53c024 was re-scheduled: Binding failed for port a78a7314-6023-4650-a7ac-583240c3f102, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 679.124458] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 679.124458] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Acquiring lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.124458] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Acquired lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.124458] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.140866] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 679.141111] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.141265] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 679.141440] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.141581] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 679.141724] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 679.141916] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 679.142410] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 679.142631] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 679.142798] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 679.142964] env[61936]: DEBUG nova.virt.hardware [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 679.144453] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7aafca6-3c62-457c-b297-97bd482ccf35 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.154028] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770615e3-dec3-42e7-ab00-6957019f3ad1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.325750] env[61936]: ERROR nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. [ 679.325750] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 679.325750] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 679.325750] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 679.325750] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.325750] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 679.325750] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.325750] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 679.325750] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.325750] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 679.325750] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.325750] env[61936]: ERROR nova.compute.manager raise self.value [ 679.325750] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.325750] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 679.325750] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.325750] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 679.328276] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.328276] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 679.328276] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. [ 679.328276] env[61936]: ERROR nova.compute.manager [ 679.328276] env[61936]: Traceback (most recent call last): [ 679.328276] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 679.328276] env[61936]: listener.cb(fileno) [ 679.328276] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.328276] env[61936]: result = function(*args, **kwargs) [ 679.328276] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.328276] env[61936]: return func(*args, **kwargs) [ 679.328276] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 679.328276] env[61936]: raise e [ 679.328276] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 679.328276] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 679.328276] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.328276] env[61936]: created_port_ids = self._update_ports_for_instance( [ 679.328276] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.328276] env[61936]: with excutils.save_and_reraise_exception(): [ 679.328276] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.328276] env[61936]: self.force_reraise() [ 679.328276] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.328276] env[61936]: raise self.value [ 679.328276] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.328276] env[61936]: updated_port = self._update_port( [ 679.328276] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.328276] env[61936]: _ensure_no_port_binding_failure(port) [ 679.328276] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.328276] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 679.329082] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. [ 679.329082] env[61936]: Removing descriptor: 19 [ 679.329082] env[61936]: ERROR nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Traceback (most recent call last): [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] yield resources [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self.driver.spawn(context, instance, image_meta, [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.329082] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] vm_ref = self.build_virtual_machine(instance, [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] for vif in network_info: [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return self._sync_wrapper(fn, *args, **kwargs) [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self.wait() [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self[:] = self._gt.wait() [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return self._exit_event.wait() [ 679.329369] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] result = hub.switch() [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return self.greenlet.switch() [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] result = function(*args, **kwargs) [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return func(*args, **kwargs) [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] raise e [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] nwinfo = self.network_api.allocate_for_instance( [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.329667] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] created_port_ids = self._update_ports_for_instance( [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] with excutils.save_and_reraise_exception(): [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self.force_reraise() [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] raise self.value [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] updated_port = self._update_port( [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] _ensure_no_port_binding_failure(port) [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.330064] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] raise exception.PortBindingFailed(port_id=port['id']) [ 679.330346] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] nova.exception.PortBindingFailed: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. [ 679.330346] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] [ 679.330346] env[61936]: INFO nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Terminating instance [ 679.429391] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.671699] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.702844] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 679.703163] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.703339] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 679.703520] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.703660] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 679.703844] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 679.704089] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 679.704282] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 679.704475] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 679.704665] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 679.704856] env[61936]: DEBUG nova.virt.hardware [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 679.705786] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd60b4d9-e100-4a59-a0c4-51235603b49d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.714124] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d438fe-9bb7-4e4f-9967-53d4dbbd0a8a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.728334] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 679.734175] env[61936]: DEBUG oslo.service.loopingcall [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.736494] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 679.736726] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7163dca9-10c1-48a5-89a2-3bd0cf211104 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.752822] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 679.752822] env[61936]: value = "task-1252732" [ 679.752822] env[61936]: _type = "Task" [ 679.752822] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.761095] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252732, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.825869] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.833367] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Acquiring lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.833537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Acquired lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.833709] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.986951] env[61936]: DEBUG nova.compute.manager [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Received event network-changed-9013d4c4-000c-4f18-ad76-45bfe0d0b3bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 679.987185] env[61936]: DEBUG nova.compute.manager [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Refreshing instance network info cache due to event network-changed-9013d4c4-000c-4f18-ad76-45bfe0d0b3bc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 679.987451] env[61936]: DEBUG oslo_concurrency.lockutils [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] Acquiring lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.266782] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252732, 'name': CreateVM_Task, 'duration_secs': 0.266575} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.267528] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 680.267528] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.267791] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.267995] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 680.270834] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1a8bc65-26ff-4f44-88e5-5c8635bd80aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.276204] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 680.276204] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a6a431-9b35-dbe7-6bcd-4eaaeeb15f4c" [ 680.276204] env[61936]: _type = "Task" [ 680.276204] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.283851] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a6a431-9b35-dbe7-6bcd-4eaaeeb15f4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.328360] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Releasing lock "refresh_cache-ecf38466-be6f-432a-bdf4-f2403e53c024" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.328687] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 680.328870] env[61936]: DEBUG nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 680.329059] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 680.358751] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.364523] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.503686] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.540640] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bf83e8-f24b-41fa-a322-196d94677858 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.548796] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4945beb-61dd-4543-80cb-429bbde27c60 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.580778] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81466735-fbcc-4f42-8110-e1c658a5b7f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.589114] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbce7334-5b45-42c0-ae2b-7557dc0fc309 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.601816] env[61936]: DEBUG nova.compute.provider_tree [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.719236] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "144ddda8-db07-4308-8fd3-4045067b5b37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.719477] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "144ddda8-db07-4308-8fd3-4045067b5b37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.789422] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a6a431-9b35-dbe7-6bcd-4eaaeeb15f4c, 'name': SearchDatastore_Task, 'duration_secs': 0.008529} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.789422] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.789422] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 680.789422] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.789560] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.789641] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 680.789902] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d1e5a0d-1f5c-47b8-971a-4888c4c4a2ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.798239] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 680.798419] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 680.799143] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd5c152e-243e-45f5-a8fa-41e0e29bb8f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.804204] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 680.804204] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527e680b-2b58-4d2b-09fa-85e87a364f5a" [ 680.804204] env[61936]: _type = "Task" [ 680.804204] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.811403] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527e680b-2b58-4d2b-09fa-85e87a364f5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.869632] env[61936]: DEBUG nova.network.neutron [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.006664] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Releasing lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.007123] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 681.007315] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.007674] env[61936]: DEBUG oslo_concurrency.lockutils [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] Acquired lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.007843] env[61936]: DEBUG nova.network.neutron [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Refreshing network info cache for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.008948] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-062c4ca5-1bbd-42a0-acb3-e77247b13cdc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.018276] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675824d4-3230-4e3f-9095-611d12582690 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.042944] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aea12163-da33-436d-ad50-00f8c4fc8ae7 could not be found. [ 681.043180] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 681.043361] env[61936]: INFO nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 681.043596] env[61936]: DEBUG oslo.service.loopingcall [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.043815] env[61936]: DEBUG nova.compute.manager [-] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 681.043908] env[61936]: DEBUG nova.network.neutron [-] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.060100] env[61936]: DEBUG nova.network.neutron [-] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.107733] env[61936]: DEBUG nova.scheduler.client.report [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 681.314936] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527e680b-2b58-4d2b-09fa-85e87a364f5a, 'name': SearchDatastore_Task, 'duration_secs': 0.008503} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.315777] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-978b9687-18e1-4de5-9775-f6b2cb3db241 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.321829] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 681.321829] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525ff897-aa6a-6a07-7dad-75690263edbd" [ 681.321829] env[61936]: _type = "Task" [ 681.321829] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.329446] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525ff897-aa6a-6a07-7dad-75690263edbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.374295] env[61936]: INFO nova.compute.manager [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] [instance: ecf38466-be6f-432a-bdf4-f2403e53c024] Took 1.05 seconds to deallocate network for instance. [ 681.530285] env[61936]: DEBUG nova.network.neutron [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.562181] env[61936]: DEBUG nova.network.neutron [-] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.613837] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.614373] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 681.617910] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.649s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.623027] env[61936]: INFO nova.compute.claims [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.653743] env[61936]: DEBUG nova.network.neutron [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.836487] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525ff897-aa6a-6a07-7dad-75690263edbd, 'name': SearchDatastore_Task, 'duration_secs': 0.008327} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.836813] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.837124] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 681.837388] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec87364f-454f-4c9b-8deb-08832e7d2cc6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.843588] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 681.843588] env[61936]: value = "task-1252733" [ 681.843588] env[61936]: _type = "Task" [ 681.843588] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.851133] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.065599] env[61936]: INFO nova.compute.manager [-] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Took 1.02 seconds to deallocate network for instance. [ 682.068335] env[61936]: DEBUG nova.compute.claims [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 682.068536] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.124481] env[61936]: DEBUG nova.compute.utils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.128574] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 682.128853] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 682.157866] env[61936]: DEBUG oslo_concurrency.lockutils [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] Releasing lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.158189] env[61936]: DEBUG nova.compute.manager [req-3900dca2-a0fc-4437-bdde-9c9b681b6946 req-b7b370ea-e0de-48d7-929c-0f4658515e1b service nova] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Received event network-vif-deleted-9013d4c4-000c-4f18-ad76-45bfe0d0b3bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 682.193414] env[61936]: DEBUG nova.policy [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '40e569098d484e258372f0c8a1aac5d2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac998c56c2234f34b840cc11aae562c7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.353975] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252733, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.414347} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.354731] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 682.355164] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 682.357706] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f8a7464-5929-45eb-8797-b2dabac2597c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.364100] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 682.364100] env[61936]: value = "task-1252734" [ 682.364100] env[61936]: _type = "Task" [ 682.364100] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.373770] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252734, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.408017] env[61936]: INFO nova.scheduler.client.report [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Deleted allocations for instance ecf38466-be6f-432a-bdf4-f2403e53c024 [ 682.629058] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 682.792405] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Successfully created port: 6e606fc5-e58d-4d6d-b757-03c0a83ebfae {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 682.873819] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252734, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057981} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.877279] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.878865] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fae915-2486-458f-9a1c-f0b7aa4e1225 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.903097] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Reconfiguring VM instance instance-00000017 to attach disk [datastore1] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.907656] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76d1ad8f-2c38-4b8d-b895-27c2944f2f04 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.922635] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0e6cb76-3703-4a75-9af5-983a2581d2b9 tempest-ImagesOneServerTestJSON-725063176 tempest-ImagesOneServerTestJSON-725063176-project-member] Lock "ecf38466-be6f-432a-bdf4-f2403e53c024" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.110s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.929357] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 682.929357] env[61936]: value = "task-1252735" [ 682.929357] env[61936]: _type = "Task" [ 682.929357] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.942655] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252735, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.043732] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192bcee9-8b93-4859-8cf0-13277209a67e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.052964] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22f5e67-bf9e-4d6a-b8cf-3060a8b02d8f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.087283] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46300f6-ddc3-4424-8e59-d4be82748de4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.095347] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd6a47c-6439-4cd1-9765-48ea7d92a6c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.111463] env[61936]: DEBUG nova.compute.provider_tree [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 683.427640] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 683.439762] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252735, 'name': ReconfigVM_Task, 'duration_secs': 0.284319} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.440321] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Reconfigured VM instance instance-00000017 to attach disk [datastore1] 994ff54f-a26b-40e0-a2c7-502f3598d9bf/994ff54f-a26b-40e0-a2c7-502f3598d9bf.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 683.441642] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-daf4b6c6-4445-4cd7-ae12-c4025dec3557 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.449120] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 683.449120] env[61936]: value = "task-1252736" [ 683.449120] env[61936]: _type = "Task" [ 683.449120] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.457663] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252736, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.637171] env[61936]: ERROR nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [req-9d383cde-5c03-4dc4-9ae9-6435f39a5d7e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID abd382a7-b427-4dcc-a050-3fb56fdd60ac. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9d383cde-5c03-4dc4-9ae9-6435f39a5d7e"}]} [ 683.641902] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 683.656239] env[61936]: DEBUG nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Refreshing inventories for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 683.667873] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 683.668128] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.668730] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 683.668931] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.669091] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 683.669237] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 683.669440] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 683.669595] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 683.669949] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 683.670135] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 683.670310] env[61936]: DEBUG nova.virt.hardware [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 683.671359] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137c5477-55e8-4414-ae5c-0979727316ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.675924] env[61936]: DEBUG nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Updating ProviderTree inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 683.676143] env[61936]: DEBUG nova.compute.provider_tree [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 683.683521] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1223308e-1025-4a00-8e90-28ec117633fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.689681] env[61936]: DEBUG nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Refreshing aggregate associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, aggregates: None {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 683.710154] env[61936]: DEBUG nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Refreshing trait associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 683.835141] env[61936]: DEBUG nova.compute.manager [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Received event network-changed-6e606fc5-e58d-4d6d-b757-03c0a83ebfae {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 683.835141] env[61936]: DEBUG nova.compute.manager [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Refreshing instance network info cache due to event network-changed-6e606fc5-e58d-4d6d-b757-03c0a83ebfae. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 683.835141] env[61936]: DEBUG oslo_concurrency.lockutils [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] Acquiring lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.835141] env[61936]: DEBUG oslo_concurrency.lockutils [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] Acquired lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.835141] env[61936]: DEBUG nova.network.neutron [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Refreshing network info cache for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 683.959021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.965802] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252736, 'name': Rename_Task, 'duration_secs': 0.184866} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.966339] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 683.967344] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c495e9d2-3f0e-46b6-9d3c-5e03ec08ac7d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.973796] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 683.973796] env[61936]: value = "task-1252737" [ 683.973796] env[61936]: _type = "Task" [ 683.973796] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.985841] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252737, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.002883] env[61936]: ERROR nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. [ 684.002883] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.002883] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.002883] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.002883] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.002883] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.002883] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.002883] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.002883] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.002883] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 684.002883] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.002883] env[61936]: ERROR nova.compute.manager raise self.value [ 684.002883] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.002883] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.002883] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.002883] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.003310] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.003310] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.003310] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. [ 684.003310] env[61936]: ERROR nova.compute.manager [ 684.003310] env[61936]: Traceback (most recent call last): [ 684.003310] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.003310] env[61936]: listener.cb(fileno) [ 684.003310] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.003310] env[61936]: result = function(*args, **kwargs) [ 684.003310] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.003310] env[61936]: return func(*args, **kwargs) [ 684.003310] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 684.003310] env[61936]: raise e [ 684.003310] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.003310] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 684.003310] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.003310] env[61936]: created_port_ids = self._update_ports_for_instance( [ 684.003310] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.003310] env[61936]: with excutils.save_and_reraise_exception(): [ 684.003310] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.003310] env[61936]: self.force_reraise() [ 684.003310] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.003310] env[61936]: raise self.value [ 684.003310] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.003310] env[61936]: updated_port = self._update_port( [ 684.003310] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.003310] env[61936]: _ensure_no_port_binding_failure(port) [ 684.003310] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.003310] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.004071] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. [ 684.004071] env[61936]: Removing descriptor: 15 [ 684.004179] env[61936]: ERROR nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Traceback (most recent call last): [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] yield resources [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self.driver.spawn(context, instance, image_meta, [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] vm_ref = self.build_virtual_machine(instance, [ 684.004179] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] for vif in network_info: [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return self._sync_wrapper(fn, *args, **kwargs) [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self.wait() [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self[:] = self._gt.wait() [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return self._exit_event.wait() [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.004776] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] result = hub.switch() [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return self.greenlet.switch() [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] result = function(*args, **kwargs) [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return func(*args, **kwargs) [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] raise e [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] nwinfo = self.network_api.allocate_for_instance( [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] created_port_ids = self._update_ports_for_instance( [ 684.005080] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] with excutils.save_and_reraise_exception(): [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self.force_reraise() [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] raise self.value [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] updated_port = self._update_port( [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] _ensure_no_port_binding_failure(port) [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] raise exception.PortBindingFailed(port_id=port['id']) [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] nova.exception.PortBindingFailed: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. [ 684.005472] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] [ 684.005828] env[61936]: INFO nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Terminating instance [ 684.153943] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0cbb41-8a83-4515-9bd4-6749cfaab3c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.161697] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49524841-45b5-44c0-beeb-1da56a44ca4a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.194107] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671af783-c1eb-4ea9-a51c-4cc604d001b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.202265] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c93872-2042-4df1-980e-8efb570b4d46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.219274] env[61936]: DEBUG nova.compute.provider_tree [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 684.381537] env[61936]: DEBUG nova.network.neutron [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.484719] env[61936]: DEBUG oslo_vmware.api [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252737, 'name': PowerOnVM_Task, 'duration_secs': 0.412537} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.485061] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 684.485309] env[61936]: DEBUG nova.compute.manager [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 684.486149] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7343fc-0bd6-4b09-9547-cbce7e185fd9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.489468] env[61936]: DEBUG nova.network.neutron [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.508380] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Acquiring lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.755642] env[61936]: DEBUG nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 56 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 684.755926] env[61936]: DEBUG nova.compute.provider_tree [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 56 to 57 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 684.756116] env[61936]: DEBUG nova.compute.provider_tree [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 684.994075] env[61936]: DEBUG oslo_concurrency.lockutils [req-fa56ebe7-7f9b-477a-8f92-9a912c9e8434 req-f8d3068d-3ade-4465-9510-58f29d5235bc service nova] Releasing lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.994315] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Acquired lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.994551] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.006770] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.261337] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.643s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.262015] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 685.264705] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.614s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.265982] env[61936]: INFO nova.compute.claims [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.510685] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.585204] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.654562] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.654912] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.655053] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.655232] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.655394] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.659496] env[61936]: INFO nova.compute.manager [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Terminating instance [ 685.771075] env[61936]: DEBUG nova.compute.utils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.776404] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 685.776588] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 685.880770] env[61936]: DEBUG nova.policy [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd228432b27541699a6916ff53981b11', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3daabe4b6f124136ae13dfe3726fd81e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.884361] env[61936]: DEBUG nova.compute.manager [req-3c1edf4f-c308-4f46-bbb0-5cacc317b8c4 req-c47de00a-ebdc-4bca-ad7d-6b042af9a3a2 service nova] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Received event network-vif-deleted-6e606fc5-e58d-4d6d-b757-03c0a83ebfae {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 686.087460] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Releasing lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.087923] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 686.088132] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 686.088449] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ddb8cd2-2f27-468b-b29d-404adfe85db4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.100202] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfa8ae5-1c44-4887-ad1d-643be30bc834 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.131404] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496 could not be found. [ 686.131404] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 686.131404] env[61936]: INFO nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Took 0.04 seconds to destroy the instance on the hypervisor. [ 686.131646] env[61936]: DEBUG oslo.service.loopingcall [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.131883] env[61936]: DEBUG nova.compute.manager [-] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 686.131976] env[61936]: DEBUG nova.network.neutron [-] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.162629] env[61936]: DEBUG nova.network.neutron [-] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.167175] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "refresh_cache-994ff54f-a26b-40e0-a2c7-502f3598d9bf" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.167175] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquired lock "refresh_cache-994ff54f-a26b-40e0-a2c7-502f3598d9bf" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.167175] env[61936]: DEBUG nova.network.neutron [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.268461] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Successfully created port: 07cd9a97-a64c-44af-958c-20bb713480e7 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.279728] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 686.667070] env[61936]: DEBUG nova.network.neutron [-] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.689956] env[61936]: DEBUG nova.network.neutron [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.713063] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5197cf-a113-4c34-832a-5ffc5f02f1ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.726384] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29881e24-d8d9-414c-b5ce-c89002e4eaf8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.757857] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25aae727-89e5-48bf-89f1-d172979baf6e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.767301] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d584b8-1f6b-4165-9353-32d695bd4150 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.787263] env[61936]: DEBUG nova.compute.provider_tree [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.794100] env[61936]: DEBUG nova.network.neutron [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.172591] env[61936]: INFO nova.compute.manager [-] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Took 1.04 seconds to deallocate network for instance. [ 687.174237] env[61936]: DEBUG nova.compute.claims [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.174635] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.291573] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 687.293139] env[61936]: DEBUG nova.scheduler.client.report [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.296888] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Releasing lock "refresh_cache-994ff54f-a26b-40e0-a2c7-502f3598d9bf" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.297383] env[61936]: DEBUG nova.compute.manager [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 687.297626] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.301022] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8f3e7b-da12-4513-8abe-a2360f19c99a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.307169] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 687.307252] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3125b946-b098-4850-96ae-003bb83a7304 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.313900] env[61936]: DEBUG oslo_vmware.api [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 687.313900] env[61936]: value = "task-1252738" [ 687.313900] env[61936]: _type = "Task" [ 687.313900] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.327891] env[61936]: DEBUG oslo_vmware.api [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252738, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.330442] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 687.331216] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.331430] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 687.331656] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.331829] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 687.332010] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 687.332264] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 687.332428] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 687.332589] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 687.332745] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 687.332910] env[61936]: DEBUG nova.virt.hardware [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 687.333837] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7150c3-3c99-474c-9d3a-53561f950dff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.341753] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e825b3-c06f-4561-adb3-684b118cd25e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.728024] env[61936]: ERROR nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. [ 687.728024] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.728024] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.728024] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.728024] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.728024] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.728024] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.728024] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.728024] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.728024] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 687.728024] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.728024] env[61936]: ERROR nova.compute.manager raise self.value [ 687.728024] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.728024] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.728024] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.728024] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.728960] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.728960] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.728960] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. [ 687.728960] env[61936]: ERROR nova.compute.manager [ 687.728960] env[61936]: Traceback (most recent call last): [ 687.728960] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.728960] env[61936]: listener.cb(fileno) [ 687.728960] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.728960] env[61936]: result = function(*args, **kwargs) [ 687.728960] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.728960] env[61936]: return func(*args, **kwargs) [ 687.728960] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.728960] env[61936]: raise e [ 687.728960] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.728960] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 687.728960] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.728960] env[61936]: created_port_ids = self._update_ports_for_instance( [ 687.728960] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.728960] env[61936]: with excutils.save_and_reraise_exception(): [ 687.728960] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.728960] env[61936]: self.force_reraise() [ 687.728960] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.728960] env[61936]: raise self.value [ 687.728960] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.728960] env[61936]: updated_port = self._update_port( [ 687.728960] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.728960] env[61936]: _ensure_no_port_binding_failure(port) [ 687.728960] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.728960] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.730103] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. [ 687.730103] env[61936]: Removing descriptor: 15 [ 687.730103] env[61936]: ERROR nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Traceback (most recent call last): [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] yield resources [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self.driver.spawn(context, instance, image_meta, [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.730103] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] vm_ref = self.build_virtual_machine(instance, [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] for vif in network_info: [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return self._sync_wrapper(fn, *args, **kwargs) [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self.wait() [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self[:] = self._gt.wait() [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return self._exit_event.wait() [ 687.730627] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] result = hub.switch() [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return self.greenlet.switch() [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] result = function(*args, **kwargs) [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return func(*args, **kwargs) [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] raise e [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] nwinfo = self.network_api.allocate_for_instance( [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.731195] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] created_port_ids = self._update_ports_for_instance( [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] with excutils.save_and_reraise_exception(): [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self.force_reraise() [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] raise self.value [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] updated_port = self._update_port( [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] _ensure_no_port_binding_failure(port) [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.731834] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] raise exception.PortBindingFailed(port_id=port['id']) [ 687.732348] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] nova.exception.PortBindingFailed: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. [ 687.732348] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] [ 687.732348] env[61936]: INFO nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Terminating instance [ 687.800961] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.801793] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 687.805964] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.490s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.824526] env[61936]: DEBUG oslo_vmware.api [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252738, 'name': PowerOffVM_Task, 'duration_secs': 0.202803} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.824765] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 687.824921] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 687.825178] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5538733-2727-4c36-af54-d26196774865 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.853743] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 687.853942] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 687.854194] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleting the datastore file [datastore1] 994ff54f-a26b-40e0-a2c7-502f3598d9bf {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 687.854508] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92b36341-a009-4e13-a5df-cbd2ae89c232 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.861360] env[61936]: DEBUG oslo_vmware.api [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for the task: (returnval){ [ 687.861360] env[61936]: value = "task-1252740" [ 687.861360] env[61936]: _type = "Task" [ 687.861360] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.872371] env[61936]: DEBUG oslo_vmware.api [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.218337] env[61936]: DEBUG nova.compute.manager [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Received event network-changed-07cd9a97-a64c-44af-958c-20bb713480e7 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 688.219263] env[61936]: DEBUG nova.compute.manager [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Refreshing instance network info cache due to event network-changed-07cd9a97-a64c-44af-958c-20bb713480e7. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 688.219447] env[61936]: DEBUG oslo_concurrency.lockutils [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] Acquiring lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.220542] env[61936]: DEBUG oslo_concurrency.lockutils [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] Acquired lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.220542] env[61936]: DEBUG nova.network.neutron [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Refreshing network info cache for port 07cd9a97-a64c-44af-958c-20bb713480e7 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 688.233629] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Acquiring lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.313259] env[61936]: DEBUG nova.compute.utils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.319878] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 688.319878] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 688.370792] env[61936]: DEBUG oslo_vmware.api [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Task: {'id': task-1252740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.105999} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.371121] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 688.371367] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 688.371577] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 688.371802] env[61936]: INFO nova.compute.manager [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Took 1.07 seconds to destroy the instance on the hypervisor. [ 688.372089] env[61936]: DEBUG oslo.service.loopingcall [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.372322] env[61936]: DEBUG nova.compute.manager [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 688.372449] env[61936]: DEBUG nova.network.neutron [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.390275] env[61936]: DEBUG nova.policy [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b05bdcace008487dba580405eb672ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90c29449db6c4beaa7f190d9225cb08a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.417425] env[61936]: DEBUG nova.network.neutron [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.555755] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "867a4bdc-8326-48b3-ba17-ca8336533c2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.556132] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.743066] env[61936]: DEBUG nova.network.neutron [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.785619] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edabdf6-36c5-4a5b-bb0d-6cd857047641 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.794077] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab9b48d-a029-4ce4-8752-5712c51eef08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.829622] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 688.834688] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92821478-a98d-4300-89ba-6e04d7f9b607 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.837798] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Successfully created port: 1f857e96-bf88-44c0-ae6e-09a4a5137e85 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.844544] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654749b8-3008-499e-9014-467a2155c522 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.862260] env[61936]: DEBUG nova.compute.provider_tree [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.919465] env[61936]: DEBUG nova.network.neutron [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.923917] env[61936]: DEBUG nova.network.neutron [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.366610] env[61936]: DEBUG nova.scheduler.client.report [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.420803] env[61936]: DEBUG oslo_concurrency.lockutils [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] Releasing lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.421111] env[61936]: DEBUG nova.compute.manager [req-12fa35a4-ecde-4456-977e-e44801eded3c req-68ecfb06-b3cd-45cb-b65d-7a4fe4982379 service nova] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Received event network-vif-deleted-07cd9a97-a64c-44af-958c-20bb713480e7 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 689.421826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Acquired lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.422017] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.425500] env[61936]: INFO nova.compute.manager [-] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Took 1.05 seconds to deallocate network for instance. [ 689.844448] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 689.872346] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 689.872572] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.872719] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 689.872889] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.873034] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 689.873173] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 689.875018] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 689.875018] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 689.875018] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 689.875018] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 689.875018] env[61936]: DEBUG nova.virt.hardware [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 689.875239] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.875320] env[61936]: ERROR nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Traceback (most recent call last): [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self.driver.spawn(context, instance, image_meta, [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] vm_ref = self.build_virtual_machine(instance, [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.875320] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] for vif in network_info: [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return self._sync_wrapper(fn, *args, **kwargs) [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self.wait() [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self[:] = self._gt.wait() [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return self._exit_event.wait() [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] result = hub.switch() [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.875646] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return self.greenlet.switch() [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] result = function(*args, **kwargs) [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] return func(*args, **kwargs) [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] raise e [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] nwinfo = self.network_api.allocate_for_instance( [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] created_port_ids = self._update_ports_for_instance( [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] with excutils.save_and_reraise_exception(): [ 689.876022] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] self.force_reraise() [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] raise self.value [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] updated_port = self._update_port( [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] _ensure_no_port_binding_failure(port) [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] raise exception.PortBindingFailed(port_id=port['id']) [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] nova.exception.PortBindingFailed: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. [ 689.876393] env[61936]: ERROR nova.compute.manager [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] [ 689.876700] env[61936]: DEBUG nova.compute.utils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 689.878223] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a99b8b-59b3-4d6a-b551-4748ef932b57 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.880954] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Build of instance ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2 was re-scheduled: Binding failed for port 38feb3b4-8ef1-4ebc-b5dd-8d024048a9c6, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 689.881392] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 689.881606] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquiring lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.881740] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Acquired lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.881962] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.883975] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.290s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.884156] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.885025] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 689.885025] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.116s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.886318] env[61936]: INFO nova.compute.claims [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.889840] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ede918-74a7-451e-976c-e7744d86920a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.901101] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb84850-7ff1-4a42-afb1-f776d2af05e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.907059] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca9ccb2-c383-413a-8401-88020aec77ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.933017] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c3991f-9983-42ed-b610-0fd14bfb4fb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.936583] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.941048] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdb05bb-0eb2-45b6-9740-4cf5fa594bca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.975026] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181509MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 689.975026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.975588] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.088897] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.091361] env[61936]: ERROR nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. [ 690.091361] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 690.091361] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 690.091361] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 690.091361] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.091361] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 690.091361] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.091361] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 690.091361] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.091361] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 690.091361] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.091361] env[61936]: ERROR nova.compute.manager raise self.value [ 690.091361] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.091361] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 690.091361] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.091361] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 690.091766] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.091766] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 690.091766] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. [ 690.091766] env[61936]: ERROR nova.compute.manager [ 690.091766] env[61936]: Traceback (most recent call last): [ 690.091766] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 690.091766] env[61936]: listener.cb(fileno) [ 690.091766] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.091766] env[61936]: result = function(*args, **kwargs) [ 690.091766] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.091766] env[61936]: return func(*args, **kwargs) [ 690.091766] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 690.091766] env[61936]: raise e [ 690.091766] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 690.091766] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 690.091766] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.091766] env[61936]: created_port_ids = self._update_ports_for_instance( [ 690.091766] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.091766] env[61936]: with excutils.save_and_reraise_exception(): [ 690.091766] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.091766] env[61936]: self.force_reraise() [ 690.091766] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.091766] env[61936]: raise self.value [ 690.091766] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.091766] env[61936]: updated_port = self._update_port( [ 690.091766] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.091766] env[61936]: _ensure_no_port_binding_failure(port) [ 690.091766] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.091766] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 690.092547] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. [ 690.092547] env[61936]: Removing descriptor: 15 [ 690.092547] env[61936]: ERROR nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] Traceback (most recent call last): [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] yield resources [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self.driver.spawn(context, instance, image_meta, [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.092547] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] vm_ref = self.build_virtual_machine(instance, [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] for vif in network_info: [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return self._sync_wrapper(fn, *args, **kwargs) [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self.wait() [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self[:] = self._gt.wait() [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return self._exit_event.wait() [ 690.092907] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] result = hub.switch() [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return self.greenlet.switch() [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] result = function(*args, **kwargs) [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return func(*args, **kwargs) [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] raise e [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] nwinfo = self.network_api.allocate_for_instance( [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.093291] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] created_port_ids = self._update_ports_for_instance( [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] with excutils.save_and_reraise_exception(): [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self.force_reraise() [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] raise self.value [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] updated_port = self._update_port( [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] _ensure_no_port_binding_failure(port) [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.093659] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] raise exception.PortBindingFailed(port_id=port['id']) [ 690.094008] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] nova.exception.PortBindingFailed: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. [ 690.094008] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] [ 690.094008] env[61936]: INFO nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Terminating instance [ 690.240743] env[61936]: DEBUG nova.compute.manager [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Received event network-changed-1f857e96-bf88-44c0-ae6e-09a4a5137e85 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 690.240948] env[61936]: DEBUG nova.compute.manager [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Refreshing instance network info cache due to event network-changed-1f857e96-bf88-44c0-ae6e-09a4a5137e85. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 690.241175] env[61936]: DEBUG oslo_concurrency.lockutils [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] Acquiring lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.241313] env[61936]: DEBUG oslo_concurrency.lockutils [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] Acquired lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.241468] env[61936]: DEBUG nova.network.neutron [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Refreshing network info cache for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 690.402816] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.494930] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.593333] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Releasing lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.593777] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 690.594029] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.594898] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0eb0ed98-19f8-489d-81da-93b06b43baf4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.597474] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.604564] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7db745-5711-45e6-b27b-f02c3ce81fa8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.627241] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c84e010f-98fc-4f40-ad8a-ec4b8cd6024a could not be found. [ 690.627476] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.627682] env[61936]: INFO nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 690.627931] env[61936]: DEBUG oslo.service.loopingcall [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.628188] env[61936]: DEBUG nova.compute.manager [-] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 690.628282] env[61936]: DEBUG nova.network.neutron [-] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.656942] env[61936]: DEBUG nova.network.neutron [-] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.778263] env[61936]: DEBUG nova.network.neutron [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.887820] env[61936]: DEBUG nova.network.neutron [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.997829] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Releasing lock "refresh_cache-ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.998056] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 690.998235] env[61936]: DEBUG nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 690.998396] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.017985] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.159698] env[61936]: DEBUG nova.network.neutron [-] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.190300] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279acab8-3f60-4d8e-902a-fe04c53b6c20 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.198154] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41229db-8300-4756-bf58-5233d073d1a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.228744] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c0541b-f143-4d73-9526-719b51f9f7c3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.235887] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ddbbd6-a354-48bf-858f-6a95d67cf03f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.248780] env[61936]: DEBUG nova.compute.provider_tree [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.390222] env[61936]: DEBUG oslo_concurrency.lockutils [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] Releasing lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.390484] env[61936]: DEBUG nova.compute.manager [req-a6a3416f-1501-41cc-a783-bbf7cb7be462 req-ab43ebaa-2fb6-4833-8b88-59255ee44248 service nova] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Received event network-vif-deleted-1f857e96-bf88-44c0-ae6e-09a4a5137e85 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 691.390822] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.390993] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 691.520823] env[61936]: DEBUG nova.network.neutron [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.664129] env[61936]: INFO nova.compute.manager [-] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Took 1.04 seconds to deallocate network for instance. [ 691.666695] env[61936]: DEBUG nova.compute.claims [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.666865] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.752907] env[61936]: DEBUG nova.scheduler.client.report [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.907216] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.005562] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.023294] env[61936]: INFO nova.compute.manager [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] [instance: ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2] Took 1.02 seconds to deallocate network for instance. [ 692.257482] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.258118] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 692.260660] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.303s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.263080] env[61936]: INFO nova.compute.claims [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.508076] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.508559] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 692.508777] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 692.509097] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d31f09c0-5dd8-4014-bb3d-cb9877d4ec0a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.518881] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9436aa69-d447-43a7-b784-e837c201640c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.544011] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a942e040-549e-4aca-8c4c-67de236e746e could not be found. [ 692.544229] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 692.544405] env[61936]: INFO nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 692.544645] env[61936]: DEBUG oslo.service.loopingcall [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.544858] env[61936]: DEBUG nova.compute.manager [-] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 692.544951] env[61936]: DEBUG nova.network.neutron [-] [instance: a942e040-549e-4aca-8c4c-67de236e746e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 692.558960] env[61936]: DEBUG nova.network.neutron [-] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.766751] env[61936]: DEBUG nova.compute.utils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.770543] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 692.772775] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 692.827315] env[61936]: DEBUG nova.policy [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1360b38c982c48ca8b2c05f10fbf51fd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0baddaffd8a249f298e748fbcee48b0b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.052288] env[61936]: INFO nova.scheduler.client.report [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Deleted allocations for instance ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2 [ 693.062305] env[61936]: DEBUG nova.network.neutron [-] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.169553] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Successfully created port: 3c386828-398a-43d1-8312-717874c6c43a {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.271978] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 693.562297] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3356b6df-1fa3-49b8-ad25-921b8382190b tempest-ServersAdminTestJSON-1871947732 tempest-ServersAdminTestJSON-1871947732-project-member] Lock "ac09ee6b-0c8c-42cd-9bbf-2ecc3927d0a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.872s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.566221] env[61936]: INFO nova.compute.manager [-] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Took 1.02 seconds to deallocate network for instance. [ 693.566221] env[61936]: DEBUG nova.compute.claims [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 693.566221] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.632141] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673268fe-5f00-4910-83b8-402fb6c5f844 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.640103] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d1d715-664e-4976-8f90-63e03e9df484 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.674254] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce407d08-261d-4ad5-b5b6-38176555ac46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.682786] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e215c2-9b84-4cce-9962-e4075477ecb2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.696765] env[61936]: DEBUG nova.compute.provider_tree [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.064443] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 694.200241] env[61936]: DEBUG nova.scheduler.client.report [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 694.285291] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 694.314278] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 694.314929] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.315242] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 694.315557] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.315815] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 694.316077] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 694.316316] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 694.316783] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 694.316970] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 694.317164] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 694.317344] env[61936]: DEBUG nova.virt.hardware [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 694.318717] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26bc8dd-264f-48f0-a878-b0b955db073e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.328475] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab50645-1514-4ec4-a779-90f4b813c175 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.476664] env[61936]: DEBUG nova.compute.manager [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Received event network-changed-3c386828-398a-43d1-8312-717874c6c43a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 694.477273] env[61936]: DEBUG nova.compute.manager [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Refreshing instance network info cache due to event network-changed-3c386828-398a-43d1-8312-717874c6c43a. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 694.477273] env[61936]: DEBUG oslo_concurrency.lockutils [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] Acquiring lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.477273] env[61936]: DEBUG oslo_concurrency.lockutils [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] Acquired lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.477608] env[61936]: DEBUG nova.network.neutron [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Refreshing network info cache for port 3c386828-398a-43d1-8312-717874c6c43a {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.576720] env[61936]: ERROR nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. [ 694.576720] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.576720] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.576720] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.576720] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.576720] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.576720] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.576720] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.576720] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.576720] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 694.576720] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.576720] env[61936]: ERROR nova.compute.manager raise self.value [ 694.576720] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.576720] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.576720] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.576720] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.577315] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.577315] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.577315] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. [ 694.577315] env[61936]: ERROR nova.compute.manager [ 694.577315] env[61936]: Traceback (most recent call last): [ 694.577315] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.577315] env[61936]: listener.cb(fileno) [ 694.577315] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.577315] env[61936]: result = function(*args, **kwargs) [ 694.577315] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.577315] env[61936]: return func(*args, **kwargs) [ 694.577315] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.577315] env[61936]: raise e [ 694.577315] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.577315] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 694.577315] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.577315] env[61936]: created_port_ids = self._update_ports_for_instance( [ 694.577315] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.577315] env[61936]: with excutils.save_and_reraise_exception(): [ 694.577315] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.577315] env[61936]: self.force_reraise() [ 694.577315] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.577315] env[61936]: raise self.value [ 694.577315] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.577315] env[61936]: updated_port = self._update_port( [ 694.577315] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.577315] env[61936]: _ensure_no_port_binding_failure(port) [ 694.577315] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.577315] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.578159] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. [ 694.578159] env[61936]: Removing descriptor: 15 [ 694.578159] env[61936]: ERROR nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Traceback (most recent call last): [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] yield resources [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self.driver.spawn(context, instance, image_meta, [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.578159] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] vm_ref = self.build_virtual_machine(instance, [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] for vif in network_info: [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return self._sync_wrapper(fn, *args, **kwargs) [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self.wait() [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self[:] = self._gt.wait() [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return self._exit_event.wait() [ 694.578529] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] result = hub.switch() [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return self.greenlet.switch() [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] result = function(*args, **kwargs) [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return func(*args, **kwargs) [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] raise e [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] nwinfo = self.network_api.allocate_for_instance( [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.578878] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] created_port_ids = self._update_ports_for_instance( [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] with excutils.save_and_reraise_exception(): [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self.force_reraise() [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] raise self.value [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] updated_port = self._update_port( [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] _ensure_no_port_binding_failure(port) [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.579261] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] raise exception.PortBindingFailed(port_id=port['id']) [ 694.579590] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] nova.exception.PortBindingFailed: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. [ 694.579590] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] [ 694.579590] env[61936]: INFO nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Terminating instance [ 694.591731] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.709027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.709027] env[61936]: DEBUG nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 694.709769] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.625s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.998033] env[61936]: DEBUG nova.network.neutron [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.083262] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Acquiring lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.101896] env[61936]: DEBUG nova.network.neutron [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.215689] env[61936]: DEBUG nova.compute.utils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.219677] env[61936]: DEBUG nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Not allocating networking since 'none' was specified. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 695.608830] env[61936]: DEBUG oslo_concurrency.lockutils [req-bf7d947a-f61d-4a54-b0dc-419ab9fbb39c req-90064c44-4e4c-43c2-a349-fa589dca0b40 service nova] Releasing lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.609596] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Acquired lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.609907] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.665081] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afafe08-458e-4ced-9015-a110fc95bd93 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.672935] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01523ac5-8b5e-4714-9c57-3453eea0490d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.706838] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a374021a-eccb-4637-9b0b-79a3165679f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.714626] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd3fdc3-690b-4378-b18e-416c238c76cc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.720098] env[61936]: DEBUG nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 695.731045] env[61936]: DEBUG nova.compute.provider_tree [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.136988] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.237271] env[61936]: DEBUG nova.scheduler.client.report [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 696.249150] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.546219] env[61936]: DEBUG nova.compute.manager [req-005815d5-9ab5-44ef-9106-2efb4342db46 req-ff7c2c00-37e8-47d7-8d1f-6dd81aa40d65 service nova] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Received event network-vif-deleted-3c386828-398a-43d1-8312-717874c6c43a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 696.741935] env[61936]: DEBUG nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 696.744632] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.746055] env[61936]: ERROR nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Traceback (most recent call last): [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self.driver.spawn(context, instance, image_meta, [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] vm_ref = self.build_virtual_machine(instance, [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.746055] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] for vif in network_info: [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return self._sync_wrapper(fn, *args, **kwargs) [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self.wait() [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self[:] = self._gt.wait() [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return self._exit_event.wait() [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] result = hub.switch() [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 696.746387] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return self.greenlet.switch() [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] result = function(*args, **kwargs) [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] return func(*args, **kwargs) [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] raise e [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] nwinfo = self.network_api.allocate_for_instance( [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] created_port_ids = self._update_ports_for_instance( [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] with excutils.save_and_reraise_exception(): [ 696.746705] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] self.force_reraise() [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] raise self.value [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] updated_port = self._update_port( [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] _ensure_no_port_binding_failure(port) [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] raise exception.PortBindingFailed(port_id=port['id']) [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] nova.exception.PortBindingFailed: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. [ 696.747020] env[61936]: ERROR nova.compute.manager [instance: 64412fc6-956f-43a6-bd5c-55745360e480] [ 696.747364] env[61936]: DEBUG nova.compute.utils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.747364] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.318s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.750268] env[61936]: INFO nova.compute.claims [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.752854] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Build of instance 64412fc6-956f-43a6-bd5c-55745360e480 was re-scheduled: Binding failed for port 3e890892-15a4-4ec1-969e-ca60c92588cc, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 696.753400] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 696.753627] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquiring lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.753981] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Acquired lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.754176] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 696.756039] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Releasing lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.758028] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 696.758028] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.758028] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-484d7143-2de4-471f-8b06-5b8f934b742a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.765783] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eccdfe4-3698-44ba-910d-468a741f0d4e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.788916] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 696.789166] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.789314] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 696.789482] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.789641] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 696.789817] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 696.790050] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 696.790221] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 696.790385] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 696.790541] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 696.790706] env[61936]: DEBUG nova.virt.hardware [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 696.791758] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa6b4f4-a702-4502-bd83-38d927e2167b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.799538] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b could not be found. [ 696.799737] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.799915] env[61936]: INFO nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 696.800166] env[61936]: DEBUG oslo.service.loopingcall [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.800748] env[61936]: DEBUG nova.compute.manager [-] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 696.800851] env[61936]: DEBUG nova.network.neutron [-] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.805989] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4e1519-d22e-43c5-b8fd-1a56165fb956 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.824291] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 696.830401] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Creating folder: Project (4b62cbc5b86e47be8bab36ee6a909c86). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 696.831861] env[61936]: DEBUG nova.network.neutron [-] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.833118] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac8db7a7-8e80-4cb6-9a2d-8ba96dd418e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.843098] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Created folder: Project (4b62cbc5b86e47be8bab36ee6a909c86) in parent group-v269874. [ 696.843360] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Creating folder: Instances. Parent ref: group-v269886. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 696.843964] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d97ad6ab-dbb0-44a8-84ee-d08f71e08729 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.852711] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Created folder: Instances in parent group-v269886. [ 696.853043] env[61936]: DEBUG oslo.service.loopingcall [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.853305] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 696.853516] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2ecdc91-cf4c-4ecd-8208-86847f365989 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.870096] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 696.870096] env[61936]: value = "task-1252743" [ 696.870096] env[61936]: _type = "Task" [ 696.870096] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.877930] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252743, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.272091] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.332795] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.337679] env[61936]: DEBUG nova.network.neutron [-] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.380040] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252743, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.843019] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Releasing lock "refresh_cache-64412fc6-956f-43a6-bd5c-55745360e480" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.843019] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 697.843019] env[61936]: DEBUG nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 697.843019] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 697.843488] env[61936]: INFO nova.compute.manager [-] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Took 1.04 seconds to deallocate network for instance. [ 697.849675] env[61936]: DEBUG nova.compute.claims [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.849894] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.868780] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.883586] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252743, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.113732] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f0695f-9798-4fce-9b02-48b4ed2145bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.123508] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda250b0-0619-40d1-b63d-813a706053fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.152613] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b253877-d013-4c13-8ef5-473593cd556b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.160272] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c37e9dc-b718-465c-b876-bfeebb6f70eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.173981] env[61936]: DEBUG nova.compute.provider_tree [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.372393] env[61936]: DEBUG nova.network.neutron [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.383029] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252743, 'name': CreateVM_Task, 'duration_secs': 1.301336} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.383806] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 698.384251] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.384407] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.384836] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 698.385251] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-808cc823-7944-4646-821b-d91ac4daa151 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.390297] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 698.390297] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c7fd6c-7679-2f28-3694-6efe59ee38ec" [ 698.390297] env[61936]: _type = "Task" [ 698.390297] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.400390] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c7fd6c-7679-2f28-3694-6efe59ee38ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.677008] env[61936]: DEBUG nova.scheduler.client.report [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 698.878842] env[61936]: INFO nova.compute.manager [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] [instance: 64412fc6-956f-43a6-bd5c-55745360e480] Took 1.04 seconds to deallocate network for instance. [ 698.904210] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c7fd6c-7679-2f28-3694-6efe59ee38ec, 'name': SearchDatastore_Task, 'duration_secs': 0.009872} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.904528] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.905073] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.905948] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.905948] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.905948] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.905948] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e53afcff-e90e-47c0-8cef-e4a30771aa15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.914201] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.914407] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 698.915151] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff35284d-ce3e-40a3-a9d2-fb85601d0c1a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.920302] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 698.920302] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5295279c-44c6-11a9-1d0f-61916d518084" [ 698.920302] env[61936]: _type = "Task" [ 698.920302] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.929552] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5295279c-44c6-11a9-1d0f-61916d518084, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.184027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.184027] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 699.191025] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.117s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.433076] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5295279c-44c6-11a9-1d0f-61916d518084, 'name': SearchDatastore_Task, 'duration_secs': 0.013665} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.434767] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-772d0bc0-a83c-425d-bfab-7a3cb98c5fb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.445020] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 699.445020] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5221c0e8-5f8e-ed57-7148-49832c5a91b7" [ 699.445020] env[61936]: _type = "Task" [ 699.445020] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.452884] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5221c0e8-5f8e-ed57-7148-49832c5a91b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.689244] env[61936]: DEBUG nova.compute.utils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.689636] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 699.692428] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.759169] env[61936]: DEBUG nova.policy [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28b6cd961e1f4de5b1e492bd69caf1c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '802e66eed7664889a984c105ab601e2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.911163] env[61936]: INFO nova.scheduler.client.report [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Deleted allocations for instance 64412fc6-956f-43a6-bd5c-55745360e480 [ 699.954798] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5221c0e8-5f8e-ed57-7148-49832c5a91b7, 'name': SearchDatastore_Task, 'duration_secs': 0.009063} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.955073] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.955327] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 699.956033] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-662f79ad-cfbf-4cca-beb6-682705394bc0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.965214] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 699.965214] env[61936]: value = "task-1252744" [ 699.965214] env[61936]: _type = "Task" [ 699.965214] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.973253] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252744, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.141403] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed886f1a-5637-4e8c-876a-3b20f4b5def8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.151945] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e9ef0c-79a1-4346-834d-9bddbad4430a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.195304] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94aa902e-a92d-4376-a8ea-0f8f97084dd3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.199154] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 700.208190] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bbfba2-0cb8-4156-b735-eda9f1ca9c9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.222726] env[61936]: DEBUG nova.compute.provider_tree [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.426965] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9378c78-ecae-4932-a636-c0b5d4d34657 tempest-DeleteServersAdminTestJSON-1806407281 tempest-DeleteServersAdminTestJSON-1806407281-project-member] Lock "64412fc6-956f-43a6-bd5c-55745360e480" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.940s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.464984] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Successfully created port: 2c1e7e34-6223-4293-989c-3cb2f28ff8f8 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.477339] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252744, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4481} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.477610] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 700.478479] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.478479] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c50decf-a96c-40e3-b4d1-9901e5e7e9f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.484620] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 700.484620] env[61936]: value = "task-1252745" [ 700.484620] env[61936]: _type = "Task" [ 700.484620] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.496056] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252745, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.732211] env[61936]: DEBUG nova.scheduler.client.report [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 700.930626] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 700.996426] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252745, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065857} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.996698] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 700.998568] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73eac2ad-6067-4e70-8329-9a9a8705e725 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.023289] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.023923] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22158acf-de4a-4b66-8ec3-db19b1c00350 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.048453] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 701.048453] env[61936]: value = "task-1252746" [ 701.048453] env[61936]: _type = "Task" [ 701.048453] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.057606] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252746, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.213157] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 701.241486] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.053s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.241486] env[61936]: ERROR nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. [ 701.241486] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Traceback (most recent call last): [ 701.241486] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 701.241486] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self.driver.spawn(context, instance, image_meta, [ 701.241486] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 701.241486] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.241486] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.241486] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] vm_ref = self.build_virtual_machine(instance, [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] for vif in network_info: [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return self._sync_wrapper(fn, *args, **kwargs) [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self.wait() [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self[:] = self._gt.wait() [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return self._exit_event.wait() [ 701.241975] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] result = hub.switch() [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return self.greenlet.switch() [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] result = function(*args, **kwargs) [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] return func(*args, **kwargs) [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] raise e [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] nwinfo = self.network_api.allocate_for_instance( [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.242598] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] created_port_ids = self._update_ports_for_instance( [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] with excutils.save_and_reraise_exception(): [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] self.force_reraise() [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] raise self.value [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] updated_port = self._update_port( [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] _ensure_no_port_binding_failure(port) [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.242991] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] raise exception.PortBindingFailed(port_id=port['id']) [ 701.243396] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] nova.exception.PortBindingFailed: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. [ 701.243396] env[61936]: ERROR nova.compute.manager [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] [ 701.243396] env[61936]: DEBUG nova.compute.utils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.243396] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.286s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.243868] env[61936]: INFO nova.compute.claims [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.246484] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Build of instance aea12163-da33-436d-ad50-00f8c4fc8ae7 was re-scheduled: Binding failed for port 9013d4c4-000c-4f18-ad76-45bfe0d0b3bc, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 701.246951] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 701.247203] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Acquiring lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.247349] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Acquired lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.247506] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.253105] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 701.253105] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.253105] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 701.253309] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.253356] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 701.254318] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 701.254318] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 701.254318] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 701.254318] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 701.254318] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 701.254659] env[61936]: DEBUG nova.virt.hardware [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 701.255426] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a72c05-16d0-4290-8dd0-14436a905f9c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.264669] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87386e46-6c68-4a74-bc2a-dcc138a4240a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.462352] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.559669] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252746, 'name': ReconfigVM_Task, 'duration_secs': 0.275199} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.559974] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 701.560581] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ed4e16a-a986-44a0-b6b6-cbc3e1e8b5ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.568145] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 701.568145] env[61936]: value = "task-1252747" [ 701.568145] env[61936]: _type = "Task" [ 701.568145] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.580592] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252747, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.642713] env[61936]: DEBUG nova.compute.manager [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Received event network-changed-2c1e7e34-6223-4293-989c-3cb2f28ff8f8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 701.642911] env[61936]: DEBUG nova.compute.manager [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Refreshing instance network info cache due to event network-changed-2c1e7e34-6223-4293-989c-3cb2f28ff8f8. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 701.643137] env[61936]: DEBUG oslo_concurrency.lockutils [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] Acquiring lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.643276] env[61936]: DEBUG oslo_concurrency.lockutils [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] Acquired lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.643433] env[61936]: DEBUG nova.network.neutron [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Refreshing network info cache for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 701.783445] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.953112] env[61936]: ERROR nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. [ 701.953112] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.953112] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.953112] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.953112] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.953112] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.953112] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.953112] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.953112] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.953112] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 701.953112] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.953112] env[61936]: ERROR nova.compute.manager raise self.value [ 701.953112] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.953112] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.953112] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.953112] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.953678] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.953678] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.953678] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. [ 701.953678] env[61936]: ERROR nova.compute.manager [ 701.953678] env[61936]: Traceback (most recent call last): [ 701.953678] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.953678] env[61936]: listener.cb(fileno) [ 701.953678] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.953678] env[61936]: result = function(*args, **kwargs) [ 701.953678] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.953678] env[61936]: return func(*args, **kwargs) [ 701.953678] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.953678] env[61936]: raise e [ 701.953678] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.953678] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 701.953678] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.953678] env[61936]: created_port_ids = self._update_ports_for_instance( [ 701.953678] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.953678] env[61936]: with excutils.save_and_reraise_exception(): [ 701.953678] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.953678] env[61936]: self.force_reraise() [ 701.953678] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.953678] env[61936]: raise self.value [ 701.953678] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.953678] env[61936]: updated_port = self._update_port( [ 701.953678] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.953678] env[61936]: _ensure_no_port_binding_failure(port) [ 701.953678] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.953678] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.954474] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. [ 701.954474] env[61936]: Removing descriptor: 15 [ 701.954474] env[61936]: ERROR nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Traceback (most recent call last): [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] yield resources [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self.driver.spawn(context, instance, image_meta, [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.954474] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] vm_ref = self.build_virtual_machine(instance, [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] for vif in network_info: [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return self._sync_wrapper(fn, *args, **kwargs) [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self.wait() [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self[:] = self._gt.wait() [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return self._exit_event.wait() [ 701.954851] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] result = hub.switch() [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return self.greenlet.switch() [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] result = function(*args, **kwargs) [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return func(*args, **kwargs) [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] raise e [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] nwinfo = self.network_api.allocate_for_instance( [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.955231] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] created_port_ids = self._update_ports_for_instance( [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] with excutils.save_and_reraise_exception(): [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self.force_reraise() [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] raise self.value [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] updated_port = self._update_port( [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] _ensure_no_port_binding_failure(port) [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.955594] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] raise exception.PortBindingFailed(port_id=port['id']) [ 701.955938] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] nova.exception.PortBindingFailed: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. [ 701.955938] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] [ 701.955938] env[61936]: INFO nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Terminating instance [ 701.996816] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.086433] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252747, 'name': Rename_Task, 'duration_secs': 0.13049} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.086550] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.086779] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e23e782-a6d2-498c-baf5-8af919dc4214 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.092973] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 702.092973] env[61936]: value = "task-1252748" [ 702.092973] env[61936]: _type = "Task" [ 702.092973] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.100555] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252748, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.173143] env[61936]: DEBUG nova.network.neutron [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.311690] env[61936]: DEBUG nova.network.neutron [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.462331] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Acquiring lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.499919] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Releasing lock "refresh_cache-aea12163-da33-436d-ad50-00f8c4fc8ae7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.500155] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 702.500408] env[61936]: DEBUG nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 702.500512] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.517568] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.602657] env[61936]: DEBUG oslo_vmware.api [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252748, 'name': PowerOnVM_Task, 'duration_secs': 0.419778} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.603034] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 702.603576] env[61936]: INFO nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Took 5.86 seconds to spawn the instance on the hypervisor. [ 702.603576] env[61936]: DEBUG nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 702.604502] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24f155c-61f4-457e-9bad-64444fb30aee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.653651] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bbbae0-5cbe-4618-b5e8-f2fad41c7f36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.661504] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84aa359-f256-4286-a0bf-471ed9045a7a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.696807] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b978e2f8-412d-4a67-bf12-d5165b0c7a97 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.706285] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2894016-c9b3-4bee-8e9f-f5e426e05948 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.718931] env[61936]: DEBUG nova.compute.provider_tree [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.779503] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "301b0504-5cf9-44e0-bd3e-342f8a89278b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.779696] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.817030] env[61936]: DEBUG oslo_concurrency.lockutils [req-80a87cbf-573a-4133-9300-6c366103536f req-478e2acf-02ce-4e2e-b846-0c9d2e72f74b service nova] Releasing lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.817432] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Acquired lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.817681] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.024142] env[61936]: DEBUG nova.network.neutron [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.153807] env[61936]: INFO nova.compute.manager [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Took 26.21 seconds to build instance. [ 703.221656] env[61936]: DEBUG nova.scheduler.client.report [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 703.339781] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.464305] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.526574] env[61936]: INFO nova.compute.manager [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] [instance: aea12163-da33-436d-ad50-00f8c4fc8ae7] Took 1.03 seconds to deallocate network for instance. [ 703.656270] env[61936]: DEBUG oslo_concurrency.lockutils [None req-49097278-002d-4be2-8526-84792c34bce2 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "1e0dade4-ea54-457d-8398-54ba810a83c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.514s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.677150] env[61936]: DEBUG nova.compute.manager [req-347bcc76-617b-4a89-948b-4e5f7e5041ce req-ca6ae7d0-511c-48cf-ae49-1ef6c37d5227 service nova] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Received event network-vif-deleted-2c1e7e34-6223-4293-989c-3cb2f28ff8f8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 703.729180] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.729180] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 703.731497] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.725s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.731741] env[61936]: DEBUG nova.objects.instance [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 703.968177] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Releasing lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.968390] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 703.968682] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.969023] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4cc0656a-d6f4-4fd2-a777-1a34abc62b9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.978382] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e766a779-4c4e-4162-af61-782f1cf464ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.001256] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 70e751e6-4109-41fd-a80f-82d22c34ea64 could not be found. [ 704.001503] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.001717] env[61936]: INFO nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Took 0.03 seconds to destroy the instance on the hypervisor. [ 704.001993] env[61936]: DEBUG oslo.service.loopingcall [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.002262] env[61936]: DEBUG nova.compute.manager [-] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 704.002368] env[61936]: DEBUG nova.network.neutron [-] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.015626] env[61936]: DEBUG nova.network.neutron [-] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.159019] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 704.236980] env[61936]: DEBUG nova.compute.utils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.242149] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 704.242149] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 704.292915] env[61936]: DEBUG nova.policy [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '093d4095d6954ef2828a7ae927c87e10', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '813fce9d06794f08a9451f6ce4226d66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.323798] env[61936]: INFO nova.compute.manager [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Rebuilding instance [ 704.365572] env[61936]: DEBUG nova.compute.manager [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 704.366550] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527dbdae-91a5-4fdd-abce-c6b2bef04904 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.519018] env[61936]: DEBUG nova.network.neutron [-] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.558014] env[61936]: INFO nova.scheduler.client.report [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Deleted allocations for instance aea12163-da33-436d-ad50-00f8c4fc8ae7 [ 704.678518] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.717444] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Successfully created port: f140be6f-84da-41d7-be72-63d83dafab5b {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.741777] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 704.744956] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5bbb498-cacd-428a-863c-c2c03a1948e3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.746291] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.572s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.021537] env[61936]: INFO nova.compute.manager [-] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Took 1.02 seconds to deallocate network for instance. [ 705.024626] env[61936]: DEBUG nova.compute.claims [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 705.024918] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.069557] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a9313ba-c7ed-40fe-a31a-f5cc7aa430ad tempest-ImagesNegativeTestJSON-577134359 tempest-ImagesNegativeTestJSON-577134359-project-member] Lock "aea12163-da33-436d-ad50-00f8c4fc8ae7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.540s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.383015] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 705.383351] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd86a15d-9356-4714-8628-2a6930dde403 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.393615] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 705.393615] env[61936]: value = "task-1252749" [ 705.393615] env[61936]: _type = "Task" [ 705.393615] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.405979] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252749, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.571867] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 705.634603] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e837d43-b7f3-4d03-b88d-6bc4a15a3aa0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.645481] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53297a43-355e-4d80-9351-4e474153181b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.674175] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c730673-ef5d-419a-8d4b-7b6febe614f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.681387] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e4973e-ee50-4bf5-9157-588d91166d02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.699428] env[61936]: DEBUG nova.compute.provider_tree [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.756023] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 705.792846] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 705.793097] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.793251] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 705.793427] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.793568] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 705.793946] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 705.793946] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 705.794064] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 705.794205] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 705.794358] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 705.794521] env[61936]: DEBUG nova.virt.hardware [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 705.795389] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9788e65d-399a-40ca-9e30-70a1b30a2ec0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.802789] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8afdc6-c61e-4b9e-b494-e1abc53cef41 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.884916] env[61936]: DEBUG nova.compute.manager [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Received event network-changed-f140be6f-84da-41d7-be72-63d83dafab5b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 705.885144] env[61936]: DEBUG nova.compute.manager [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Refreshing instance network info cache due to event network-changed-f140be6f-84da-41d7-be72-63d83dafab5b. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 705.885371] env[61936]: DEBUG oslo_concurrency.lockutils [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] Acquiring lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.885536] env[61936]: DEBUG oslo_concurrency.lockutils [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] Acquired lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.885724] env[61936]: DEBUG nova.network.neutron [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Refreshing network info cache for port f140be6f-84da-41d7-be72-63d83dafab5b {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 705.902967] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252749, 'name': PowerOffVM_Task, 'duration_secs': 0.10731} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.903344] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 705.903569] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.904319] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89f993d-e7fc-4938-9ce7-8e68a00c32f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.914230] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 705.914230] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00cda724-b66b-4a25-b087-ce47e4ad7897 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.939190] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 705.939190] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 705.939190] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Deleting the datastore file [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 705.939190] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bffd42bb-bae7-4893-a302-798aa0014392 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.947108] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 705.947108] env[61936]: value = "task-1252751" [ 705.947108] env[61936]: _type = "Task" [ 705.947108] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.952996] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252751, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.104358] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.108353] env[61936]: ERROR nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. [ 706.108353] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.108353] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 706.108353] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.108353] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.108353] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.108353] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.108353] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.108353] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.108353] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 706.108353] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.108353] env[61936]: ERROR nova.compute.manager raise self.value [ 706.108353] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.108353] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.108353] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.108353] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.109117] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.109117] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.109117] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. [ 706.109117] env[61936]: ERROR nova.compute.manager [ 706.109117] env[61936]: Traceback (most recent call last): [ 706.109117] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.109117] env[61936]: listener.cb(fileno) [ 706.109117] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.109117] env[61936]: result = function(*args, **kwargs) [ 706.109117] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.109117] env[61936]: return func(*args, **kwargs) [ 706.109117] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 706.109117] env[61936]: raise e [ 706.109117] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 706.109117] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 706.109117] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.109117] env[61936]: created_port_ids = self._update_ports_for_instance( [ 706.109117] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.109117] env[61936]: with excutils.save_and_reraise_exception(): [ 706.109117] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.109117] env[61936]: self.force_reraise() [ 706.109117] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.109117] env[61936]: raise self.value [ 706.109117] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.109117] env[61936]: updated_port = self._update_port( [ 706.109117] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.109117] env[61936]: _ensure_no_port_binding_failure(port) [ 706.109117] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.109117] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.110735] env[61936]: nova.exception.PortBindingFailed: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. [ 706.110735] env[61936]: Removing descriptor: 19 [ 706.110735] env[61936]: ERROR nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Traceback (most recent call last): [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] yield resources [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self.driver.spawn(context, instance, image_meta, [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.110735] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] vm_ref = self.build_virtual_machine(instance, [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] for vif in network_info: [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return self._sync_wrapper(fn, *args, **kwargs) [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self.wait() [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self[:] = self._gt.wait() [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return self._exit_event.wait() [ 706.111187] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] result = hub.switch() [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return self.greenlet.switch() [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] result = function(*args, **kwargs) [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return func(*args, **kwargs) [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] raise e [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] nwinfo = self.network_api.allocate_for_instance( [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.114432] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] created_port_ids = self._update_ports_for_instance( [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] with excutils.save_and_reraise_exception(): [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self.force_reraise() [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] raise self.value [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] updated_port = self._update_port( [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] _ensure_no_port_binding_failure(port) [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.114900] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] raise exception.PortBindingFailed(port_id=port['id']) [ 706.115451] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] nova.exception.PortBindingFailed: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. [ 706.115451] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] [ 706.115451] env[61936]: INFO nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Terminating instance [ 706.206029] env[61936]: DEBUG nova.scheduler.client.report [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 706.409411] env[61936]: DEBUG nova.network.neutron [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.455057] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252751, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094339} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.455344] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 706.455553] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 706.455760] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.507175] env[61936]: DEBUG nova.network.neutron [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.613818] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Acquiring lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.709120] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.709584] env[61936]: ERROR nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Traceback (most recent call last): [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self.driver.spawn(context, instance, image_meta, [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] vm_ref = self.build_virtual_machine(instance, [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.709584] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] for vif in network_info: [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return self._sync_wrapper(fn, *args, **kwargs) [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self.wait() [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self[:] = self._gt.wait() [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return self._exit_event.wait() [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] result = hub.switch() [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.709897] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return self.greenlet.switch() [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] result = function(*args, **kwargs) [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] return func(*args, **kwargs) [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] raise e [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] nwinfo = self.network_api.allocate_for_instance( [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] created_port_ids = self._update_ports_for_instance( [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] with excutils.save_and_reraise_exception(): [ 706.710369] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] self.force_reraise() [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] raise self.value [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] updated_port = self._update_port( [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] _ensure_no_port_binding_failure(port) [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] raise exception.PortBindingFailed(port_id=port['id']) [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] nova.exception.PortBindingFailed: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. [ 706.710723] env[61936]: ERROR nova.compute.manager [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] [ 706.711064] env[61936]: DEBUG nova.compute.utils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 706.711765] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.775s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.711995] env[61936]: DEBUG nova.objects.instance [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lazy-loading 'resources' on Instance uuid 994ff54f-a26b-40e0-a2c7-502f3598d9bf {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 706.713396] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Build of instance df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496 was re-scheduled: Binding failed for port 6e606fc5-e58d-4d6d-b757-03c0a83ebfae, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 706.713833] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 706.714152] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Acquiring lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.714196] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Acquired lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.714346] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.010016] env[61936]: DEBUG oslo_concurrency.lockutils [req-f9f916ed-8702-402f-b5f5-9a3eb583ecf9 req-303c9c0e-2bfe-4db1-9305-cba8191074a7 service nova] Releasing lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.010689] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Acquired lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.011131] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.252836] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.372039] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.505762] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 707.506261] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.506510] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 707.506918] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.507190] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 707.507443] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 707.508283] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 707.508590] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 707.510534] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 707.510534] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 707.510534] env[61936]: DEBUG nova.virt.hardware [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 707.514336] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cde26ab-d1fa-4455-abf3-f768bac2b5c3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.526884] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ae46a3-7dba-4275-9a1c-d00206af36f0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.542221] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 707.547142] env[61936]: DEBUG oslo.service.loopingcall [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.547974] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.551667] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 707.552142] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1126fd2a-8cc4-4cb0-a4fa-969679c2a434 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.570716] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 707.570716] env[61936]: value = "task-1252752" [ 707.570716] env[61936]: _type = "Task" [ 707.570716] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.580160] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252752, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.667095] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabf32c9-d5a6-414e-8a01-a6b26691f889 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.675083] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4246bdf-3d0e-4512-927b-8920e257f77b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.707025] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.709969] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043ea672-d379-4862-a520-726c8f471b1d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.717430] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ea31c5-393a-48a3-8694-92edb9291438 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.732801] env[61936]: DEBUG nova.compute.provider_tree [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 707.874461] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Releasing lock "refresh_cache-df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.875104] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 707.875104] env[61936]: DEBUG nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 707.875104] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.892166] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.914173] env[61936]: DEBUG nova.compute.manager [req-e7e30a14-386e-4d54-96c9-38d2fbd3035b req-50695a4d-db57-49b5-8a8d-850d39f1c41f service nova] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Received event network-vif-deleted-f140be6f-84da-41d7-be72-63d83dafab5b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 708.080857] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252752, 'name': CreateVM_Task, 'duration_secs': 0.236297} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.081039] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 708.081454] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.081612] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.081979] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 708.082245] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32ac4e0a-02b9-42ba-80eb-da9d964c4127 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.088873] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 708.088873] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a5ef95-63d4-ae0e-085d-30792ffe1fda" [ 708.088873] env[61936]: _type = "Task" [ 708.088873] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.096566] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a5ef95-63d4-ae0e-085d-30792ffe1fda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.213426] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Releasing lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.213848] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 708.214049] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 708.214404] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b72e2b67-fd9a-47bf-8563-242baf2f9481 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.225335] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472efa03-3ecd-4a8b-b801-e9c81e5f2373 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.262758] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 64e78c61-85f4-4914-8fab-3ced64335275 could not be found. [ 708.264807] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.264807] env[61936]: INFO nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Took 0.05 seconds to destroy the instance on the hypervisor. [ 708.264807] env[61936]: DEBUG oslo.service.loopingcall [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.264807] env[61936]: DEBUG nova.compute.manager [-] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 708.264807] env[61936]: DEBUG nova.network.neutron [-] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.269236] env[61936]: ERROR nova.scheduler.client.report [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] [req-d79d5e40-48d2-4ef2-b6a6-3ac90b8a9304] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID abd382a7-b427-4dcc-a050-3fb56fdd60ac. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d79d5e40-48d2-4ef2-b6a6-3ac90b8a9304"}]} [ 708.282190] env[61936]: DEBUG nova.network.neutron [-] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.289296] env[61936]: DEBUG nova.scheduler.client.report [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Refreshing inventories for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 708.307913] env[61936]: DEBUG nova.scheduler.client.report [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Updating ProviderTree inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 708.308154] env[61936]: DEBUG nova.compute.provider_tree [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.319278] env[61936]: DEBUG nova.scheduler.client.report [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Refreshing aggregate associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, aggregates: None {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 708.339066] env[61936]: DEBUG nova.scheduler.client.report [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Refreshing trait associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 708.395319] env[61936]: DEBUG nova.network.neutron [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.602188] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a5ef95-63d4-ae0e-085d-30792ffe1fda, 'name': SearchDatastore_Task, 'duration_secs': 0.008532} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.604542] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.604784] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 708.605017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.605170] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.605391] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 708.605772] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-514ddf7d-18ae-4b8e-bb9f-cf7ffea7162b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.613601] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 708.613777] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 708.614504] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3478e88e-aeed-43ed-8738-9538328f8e34 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.621663] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 708.621663] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a04621-ea0f-a2fe-a3cb-6b461a8cfead" [ 708.621663] env[61936]: _type = "Task" [ 708.621663] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.629741] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a04621-ea0f-a2fe-a3cb-6b461a8cfead, 'name': SearchDatastore_Task, 'duration_secs': 0.00772} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.632738] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ec62b8-dd5d-40c5-a76e-e0715234bf2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.637444] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 708.637444] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b1fc0-62c7-a01a-0be1-6092c2920ee8" [ 708.637444] env[61936]: _type = "Task" [ 708.637444] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.644597] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b1fc0-62c7-a01a-0be1-6092c2920ee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.689654] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be6a716-97af-45b5-bf57-0820be78611a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.696714] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d50658e-686e-4ac6-a5d0-5fe3ff17fc15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.727013] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5087b1-e256-4136-9825-616ee9e2db4a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.734330] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33edaf66-c4fe-4d40-b7bc-33d5039202aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.749816] env[61936]: DEBUG nova.compute.provider_tree [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.784708] env[61936]: DEBUG nova.network.neutron [-] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.898710] env[61936]: INFO nova.compute.manager [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] [instance: df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496] Took 1.02 seconds to deallocate network for instance. [ 709.151135] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b1fc0-62c7-a01a-0be1-6092c2920ee8, 'name': SearchDatastore_Task, 'duration_secs': 0.008732} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.151408] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.152059] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 709.152339] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c577172-fafc-4d31-9e3b-094e9fb1f7bc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.159974] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 709.159974] env[61936]: value = "task-1252753" [ 709.159974] env[61936]: _type = "Task" [ 709.159974] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.166918] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252753, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.284539] env[61936]: DEBUG nova.scheduler.client.report [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 59 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 709.284800] env[61936]: DEBUG nova.compute.provider_tree [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 59 to 60 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 709.284979] env[61936]: DEBUG nova.compute.provider_tree [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 709.288749] env[61936]: INFO nova.compute.manager [-] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Took 1.02 seconds to deallocate network for instance. [ 709.290916] env[61936]: DEBUG nova.compute.claims [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.291115] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.670059] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252753, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461034} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.670395] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 709.670540] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.670833] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74e2339d-0587-44ac-ab1b-7db853db891d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.676992] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 709.676992] env[61936]: value = "task-1252754" [ 709.676992] env[61936]: _type = "Task" [ 709.676992] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.685265] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252754, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.790279] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.078s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.792584] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.818s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.827996] env[61936]: INFO nova.scheduler.client.report [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Deleted allocations for instance 994ff54f-a26b-40e0-a2c7-502f3598d9bf [ 709.928770] env[61936]: INFO nova.scheduler.client.report [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Deleted allocations for instance df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496 [ 710.187702] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06483} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.188013] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.188924] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4911339d-5708-4c2c-95eb-4a898e67323b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.208380] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.208635] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-507bc03d-9107-4002-85e6-33270d1acb92 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.230148] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 710.230148] env[61936]: value = "task-1252755" [ 710.230148] env[61936]: _type = "Task" [ 710.230148] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.237634] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252755, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.332576] env[61936]: DEBUG oslo_concurrency.lockutils [None req-488f27ec-b8f0-4eea-af63-06d9739edfc3 tempest-ServerShowV257Test-1916958628 tempest-ServerShowV257Test-1916958628-project-member] Lock "994ff54f-a26b-40e0-a2c7-502f3598d9bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.678s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.439054] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e825c792-2eb7-422e-a5fc-1821a579a907 tempest-ImagesOneServerNegativeTestJSON-501115881 tempest-ImagesOneServerNegativeTestJSON-501115881-project-member] Lock "df59e5e8-aeb8-4df9-a2a7-fd0c1cf11496" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.959s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.745931] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252755, 'name': ReconfigVM_Task, 'duration_secs': 0.253972} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.747198] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.747900] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5e89526-e3fc-4c31-9a17-98cdc4199336 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.756538] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 710.756538] env[61936]: value = "task-1252756" [ 710.756538] env[61936]: _type = "Task" [ 710.756538] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.765453] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252756, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.842907] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance c84e010f-98fc-4f40-ad8a-ec4b8cd6024a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 710.842907] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance a942e040-549e-4aca-8c4c-67de236e746e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 710.842907] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 710.842907] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 1e0dade4-ea54-457d-8398-54ba810a83c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 710.843336] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 70e751e6-4109-41fd-a80f-82d22c34ea64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 710.843336] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 64e78c61-85f4-4914-8fab-3ced64335275 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 710.945251] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 711.268604] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252756, 'name': Rename_Task, 'duration_secs': 0.131765} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.269083] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 711.269389] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebabe34f-98f1-485d-9401-fb010adb4637 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.276513] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 711.276513] env[61936]: value = "task-1252757" [ 711.276513] env[61936]: _type = "Task" [ 711.276513] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.284502] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252757, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.345603] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 9d805bf8-f2cd-4108-9007-2de18e1bf283 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 711.472986] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.789038] env[61936]: DEBUG oslo_vmware.api [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252757, 'name': PowerOnVM_Task, 'duration_secs': 0.427374} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.789038] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 711.789038] env[61936]: DEBUG nova.compute.manager [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 711.789038] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b1b592-842f-4861-b0d8-6cb200762d9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.849661] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 66e52362-2f67-4e5a-823a-eed7a87d1a1d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 711.900452] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.900452] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.302280] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.354338] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance a2bb2a06-0fb8-422c-bd16-8f369e397dd2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.857878] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance bdf4c76e-2441-4f76-bf74-2718c57c13e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 713.361997] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f6c3575f-9351-4957-a200-4e4d0557b55b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 713.531908] env[61936]: INFO nova.compute.manager [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Rebuilding instance [ 713.578331] env[61936]: DEBUG nova.compute.manager [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 713.579236] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1e2137-4d7d-4578-b1ba-251267181e72 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.866383] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance d7599598-492e-4d42-8417-960df12b2cd7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 714.370813] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 201d06f9-d277-4d8e-84ee-84e1053c1137 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 714.592972] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 714.593311] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76c19caf-8ec8-496b-8b60-40338c4722d7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.601966] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 714.601966] env[61936]: value = "task-1252758" [ 714.601966] env[61936]: _type = "Task" [ 714.601966] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.613316] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252758, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.874887] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 03ab71e9-64b7-4534-a693-473f3903c511 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 715.111411] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252758, 'name': PowerOffVM_Task, 'duration_secs': 0.113445} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.111661] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 715.111879] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.112633] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a259ab5b-98a8-4c7b-86ac-a4df2b08481b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.118677] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 715.118880] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac9efdd0-f57f-47f4-b310-9fb4ae5acf27 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.142497] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 715.142698] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 715.142870] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Deleting the datastore file [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 715.143113] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f3cc3e0-358b-4521-b852-9c70e5980938 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.149105] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 715.149105] env[61936]: value = "task-1252760" [ 715.149105] env[61936]: _type = "Task" [ 715.149105] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.155949] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252760, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.378317] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 0b14583c-dd0f-429d-a386-2acda353465a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 715.658975] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252760, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085728} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.659245] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 715.659424] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 715.659593] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 715.881415] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 716.384561] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 498a77d6-ca0c-4841-b4a2-b7a024281c6a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 716.695285] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 716.695692] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.695944] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 716.696254] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.696434] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 716.696595] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 716.696779] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 716.696931] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 716.697101] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 716.697262] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 716.697428] env[61936]: DEBUG nova.virt.hardware [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 716.698317] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefa8222-6269-4216-8004-253288b1002e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.706102] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716fa460-7fe6-4f03-8b48-72fe1689d47e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.719528] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.725094] env[61936]: DEBUG oslo.service.loopingcall [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.725311] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 716.725501] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8e18414-ba74-480c-bcad-d84637ccce1c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.743090] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.743090] env[61936]: value = "task-1252761" [ 716.743090] env[61936]: _type = "Task" [ 716.743090] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.750091] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252761, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.888241] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 717.252610] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252761, 'name': CreateVM_Task, 'duration_secs': 0.232095} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.254021] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 717.254021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.254021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.254021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 717.254264] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce6b8297-fea2-4605-8ba9-acb920f5a671 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.258325] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 717.258325] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523cda2a-7940-83ac-e654-6c959aa0dae8" [ 717.258325] env[61936]: _type = "Task" [ 717.258325] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.265543] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523cda2a-7940-83ac-e654-6c959aa0dae8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.392772] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 4755f16d-254b-47e9-8b50-05df95690445 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 717.769204] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523cda2a-7940-83ac-e654-6c959aa0dae8, 'name': SearchDatastore_Task, 'duration_secs': 0.011347} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.769479] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.769708] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.769942] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.770114] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.770294] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.770543] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db6028ad-ba81-4704-90ac-39925703658d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.778104] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.778274] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 717.778936] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca86e35d-3f1b-4313-a110-bc1f1259c447 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.783593] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 717.783593] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5287f787-9e75-e8f8-aff0-deebedd56081" [ 717.783593] env[61936]: _type = "Task" [ 717.783593] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.790312] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5287f787-9e75-e8f8-aff0-deebedd56081, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.894482] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 2fdb605e-72d5-4eef-bab2-0601e730243a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 718.294203] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5287f787-9e75-e8f8-aff0-deebedd56081, 'name': SearchDatastore_Task, 'duration_secs': 0.007863} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.294979] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-991619f1-9d1f-41e9-9f77-9b27945a2b8b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.300222] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 718.300222] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5289b68f-3b37-de3b-2a6a-985ad48f6931" [ 718.300222] env[61936]: _type = "Task" [ 718.300222] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.307489] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5289b68f-3b37-de3b-2a6a-985ad48f6931, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.397546] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 50cb3f79-39b2-46ab-a35c-fdad3bb266cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 718.810609] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5289b68f-3b37-de3b-2a6a-985ad48f6931, 'name': SearchDatastore_Task, 'duration_secs': 0.028747} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.810921] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.811199] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 718.811452] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0855fbcd-e430-42c6-aa0b-5d64a7df03a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.817755] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 718.817755] env[61936]: value = "task-1252762" [ 718.817755] env[61936]: _type = "Task" [ 718.817755] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.825850] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252762, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.901151] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 1251e7f5-684c-428f-9d63-60b77084d09e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 719.328023] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252762, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447949} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.328023] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 719.328267] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 719.328417] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2c46747-f94a-40b4-98f4-93ae4a81759b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.334929] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 719.334929] env[61936]: value = "task-1252763" [ 719.334929] env[61936]: _type = "Task" [ 719.334929] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.341871] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.406827] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 144ddda8-db07-4308-8fd3-4045067b5b37 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 719.845144] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055772} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.845426] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.846248] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cdca75-77ab-41d3-a512-5a84a91be7b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.865856] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.866184] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-136081a3-aa23-4825-87c6-4d70b68c81e9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.886089] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 719.886089] env[61936]: value = "task-1252764" [ 719.886089] env[61936]: _type = "Task" [ 719.886089] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.894232] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252764, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.910131] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 867a4bdc-8326-48b3-ba17-ca8336533c2d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 720.397430] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252764, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.413429] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 301b0504-5cf9-44e0-bd3e-342f8a89278b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 720.413680] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 720.413826] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 720.775697] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c925086f-8684-4224-abe1-294769e8dfaf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.783311] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd56c957-e8e4-47f6-9843-3e0191f053c8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.812974] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092b4020-ea8c-489a-aaba-bd64bcb102ee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.820478] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4fc953-3551-494f-bb2e-44f51c125719 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.833811] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 720.897582] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252764, 'name': ReconfigVM_Task, 'duration_secs': 0.88741} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.897834] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2/1e0dade4-ea54-457d-8398-54ba810a83c2.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.898568] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e3182d3-877b-46cf-8d6a-8b0c0686ccc4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.904802] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 720.904802] env[61936]: value = "task-1252765" [ 720.904802] env[61936]: _type = "Task" [ 720.904802] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.913016] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252765, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.353238] env[61936]: ERROR nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [req-439d398f-b73f-4efa-9aed-6e134fa803be] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID abd382a7-b427-4dcc-a050-3fb56fdd60ac. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-439d398f-b73f-4efa-9aed-6e134fa803be"}]} [ 721.370052] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Refreshing inventories for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 721.382654] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Updating ProviderTree inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 721.382848] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.393237] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Refreshing aggregate associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, aggregates: None {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 721.409900] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Refreshing trait associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 721.414756] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252765, 'name': Rename_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.674578] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3944ef-f6e1-402a-837f-750e2ad96d69 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.682252] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66354674-d5ec-4f22-981f-de739fefd133 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.712877] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692acc45-3bad-4fd7-8ac4-bf4aa3133137 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.719956] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d793f03-db71-403a-a989-549db80fc777 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.732841] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.915444] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252765, 'name': Rename_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.264263] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 61 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 722.265131] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 61 to 62 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 722.265131] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.416536] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252765, 'name': Rename_Task, 'duration_secs': 1.138417} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.416536] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 722.416536] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51ab98d9-84b6-480e-9c54-f64b62ad76b3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.423297] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Waiting for the task: (returnval){ [ 722.423297] env[61936]: value = "task-1252766" [ 722.423297] env[61936]: _type = "Task" [ 722.423297] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.430728] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252766, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.769478] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 722.769726] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.977s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.770017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.103s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.934067] env[61936]: DEBUG oslo_vmware.api [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Task: {'id': task-1252766, 'name': PowerOnVM_Task, 'duration_secs': 0.408896} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.934067] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 722.934067] env[61936]: DEBUG nova.compute.manager [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 722.934618] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2979a8c3-6449-456a-a221-39c9701b41a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.453993] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.574797] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe50dde-c998-4e00-9889-576c09aaaf4f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.583401] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cf6f8c-f79d-4b0d-ac57-5c2a54b5cce0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.612226] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4332467d-c7dd-4666-a107-62b782c3e52e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.618825] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42555151-1592-4656-9ab6-27e4a10593f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.631246] env[61936]: DEBUG nova.compute.provider_tree [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.134716] env[61936]: DEBUG nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 724.489715] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "1e0dade4-ea54-457d-8398-54ba810a83c2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.490063] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "1e0dade4-ea54-457d-8398-54ba810a83c2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.490297] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "1e0dade4-ea54-457d-8398-54ba810a83c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.490477] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "1e0dade4-ea54-457d-8398-54ba810a83c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.490650] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "1e0dade4-ea54-457d-8398-54ba810a83c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.492986] env[61936]: INFO nova.compute.manager [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Terminating instance [ 724.639852] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.640547] env[61936]: ERROR nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Traceback (most recent call last): [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self.driver.spawn(context, instance, image_meta, [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] vm_ref = self.build_virtual_machine(instance, [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.640547] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] for vif in network_info: [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return self._sync_wrapper(fn, *args, **kwargs) [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self.wait() [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self[:] = self._gt.wait() [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return self._exit_event.wait() [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] result = hub.switch() [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 724.640939] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return self.greenlet.switch() [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] result = function(*args, **kwargs) [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] return func(*args, **kwargs) [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] raise e [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] nwinfo = self.network_api.allocate_for_instance( [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] created_port_ids = self._update_ports_for_instance( [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] with excutils.save_and_reraise_exception(): [ 724.641407] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] self.force_reraise() [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] raise self.value [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] updated_port = self._update_port( [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] _ensure_no_port_binding_failure(port) [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] raise exception.PortBindingFailed(port_id=port['id']) [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] nova.exception.PortBindingFailed: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. [ 724.641713] env[61936]: ERROR nova.compute.manager [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] [ 724.641977] env[61936]: DEBUG nova.compute.utils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.642568] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.077s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.645663] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Build of instance c84e010f-98fc-4f40-ad8a-ec4b8cd6024a was re-scheduled: Binding failed for port 07cd9a97-a64c-44af-958c-20bb713480e7, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 724.646219] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 724.646537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Acquiring lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.646719] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Acquired lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.646883] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.996828] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "refresh_cache-1e0dade4-ea54-457d-8398-54ba810a83c2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.997020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquired lock "refresh_cache-1e0dade4-ea54-457d-8398-54ba810a83c2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.997205] env[61936]: DEBUG nova.network.neutron [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.163827] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.240075] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.451137] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a622d6c-fa87-450f-b0d1-39d7d6fedff6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.458739] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c6af25-5fd3-4678-aeb6-2c9e841f1eab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.488927] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac569eed-876a-4c75-8b61-6db5ae462e52 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.496273] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3841859-9411-40c6-971c-57b251ed6dd6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.511670] env[61936]: DEBUG nova.compute.provider_tree [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.514427] env[61936]: DEBUG nova.network.neutron [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.579882] env[61936]: DEBUG nova.network.neutron [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.745100] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Releasing lock "refresh_cache-c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.745381] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 725.745574] env[61936]: DEBUG nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 725.746011] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.762488] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.016212] env[61936]: DEBUG nova.scheduler.client.report [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 726.083318] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Releasing lock "refresh_cache-1e0dade4-ea54-457d-8398-54ba810a83c2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.083659] env[61936]: DEBUG nova.compute.manager [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 726.083859] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.084715] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbf669c-5e31-4f53-b6d4-719e8227fed4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.092388] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 726.092601] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efc6cb24-adf9-4d80-8564-8ef48273ed30 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.098357] env[61936]: DEBUG oslo_vmware.api [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 726.098357] env[61936]: value = "task-1252767" [ 726.098357] env[61936]: _type = "Task" [ 726.098357] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.105388] env[61936]: DEBUG oslo_vmware.api [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252767, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.268994] env[61936]: DEBUG nova.network.neutron [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.520588] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.878s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.521258] env[61936]: ERROR nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] Traceback (most recent call last): [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self.driver.spawn(context, instance, image_meta, [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] vm_ref = self.build_virtual_machine(instance, [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.521258] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] for vif in network_info: [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return self._sync_wrapper(fn, *args, **kwargs) [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self.wait() [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self[:] = self._gt.wait() [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return self._exit_event.wait() [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] result = hub.switch() [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.521567] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return self.greenlet.switch() [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] result = function(*args, **kwargs) [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] return func(*args, **kwargs) [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] raise e [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] nwinfo = self.network_api.allocate_for_instance( [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] created_port_ids = self._update_ports_for_instance( [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] with excutils.save_and_reraise_exception(): [ 726.521882] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] self.force_reraise() [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] raise self.value [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] updated_port = self._update_port( [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] _ensure_no_port_binding_failure(port) [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] raise exception.PortBindingFailed(port_id=port['id']) [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] nova.exception.PortBindingFailed: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. [ 726.522251] env[61936]: ERROR nova.compute.manager [instance: a942e040-549e-4aca-8c4c-67de236e746e] [ 726.522827] env[61936]: DEBUG nova.compute.utils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.523348] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.932s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.525345] env[61936]: INFO nova.compute.claims [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.529026] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Build of instance a942e040-549e-4aca-8c4c-67de236e746e was re-scheduled: Binding failed for port 1f857e96-bf88-44c0-ae6e-09a4a5137e85, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 726.529575] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 726.529798] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.529944] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.530109] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.608297] env[61936]: DEBUG oslo_vmware.api [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252767, 'name': PowerOffVM_Task, 'duration_secs': 0.112638} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.608531] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 726.608717] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 726.608955] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee7c74c6-c9c1-4304-bd9c-e8b0878c4219 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.632800] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 726.633047] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 726.633235] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Deleting the datastore file [datastore1] 1e0dade4-ea54-457d-8398-54ba810a83c2 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 726.633488] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be76c8a5-184f-4650-a500-105eff6fc88d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.640462] env[61936]: DEBUG oslo_vmware.api [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for the task: (returnval){ [ 726.640462] env[61936]: value = "task-1252769" [ 726.640462] env[61936]: _type = "Task" [ 726.640462] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.648689] env[61936]: DEBUG oslo_vmware.api [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252769, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.772182] env[61936]: INFO nova.compute.manager [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] [instance: c84e010f-98fc-4f40-ad8a-ec4b8cd6024a] Took 1.03 seconds to deallocate network for instance. [ 727.049741] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.139747] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.149528] env[61936]: DEBUG oslo_vmware.api [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Task: {'id': task-1252769, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085474} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.150283] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 727.150479] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 727.150632] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.150800] env[61936]: INFO nova.compute.manager [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Took 1.07 seconds to destroy the instance on the hypervisor. [ 727.151035] env[61936]: DEBUG oslo.service.loopingcall [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.151219] env[61936]: DEBUG nova.compute.manager [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 727.151310] env[61936]: DEBUG nova.network.neutron [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.164447] env[61936]: DEBUG nova.network.neutron [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.645246] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-a942e040-549e-4aca-8c4c-67de236e746e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.645485] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 727.645617] env[61936]: DEBUG nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 727.645841] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.666508] env[61936]: DEBUG nova.network.neutron [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.673871] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.799042] env[61936]: INFO nova.scheduler.client.report [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Deleted allocations for instance c84e010f-98fc-4f40-ad8a-ec4b8cd6024a [ 727.860532] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7835d68e-8d34-47d8-b883-df2ad1619bce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.867897] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe8f4c7-153d-4256-ade2-43fb32a6a6c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.899293] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ca19e5-2cb7-40a1-930e-300310fe713d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.906537] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153ce9fd-2d88-4970-8024-c9375764c802 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.919384] env[61936]: DEBUG nova.compute.provider_tree [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 728.169590] env[61936]: INFO nova.compute.manager [-] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Took 1.02 seconds to deallocate network for instance. [ 728.177046] env[61936]: DEBUG nova.network.neutron [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.311673] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aeff82e9-516b-4bcf-8d43-4ff3d0ac7c61 tempest-ServerPasswordTestJSON-141872129 tempest-ServerPasswordTestJSON-141872129-project-member] Lock "c84e010f-98fc-4f40-ad8a-ec4b8cd6024a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.814s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.448102] env[61936]: DEBUG nova.scheduler.client.report [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 62 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 728.449071] env[61936]: DEBUG nova.compute.provider_tree [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 62 to 63 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 728.449071] env[61936]: DEBUG nova.compute.provider_tree [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 728.676548] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.679206] env[61936]: INFO nova.compute.manager [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: a942e040-549e-4aca-8c4c-67de236e746e] Took 1.03 seconds to deallocate network for instance. [ 728.814234] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 728.957574] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.958180] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 728.960904] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.111s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.336159] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.467500] env[61936]: DEBUG nova.compute.utils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.472940] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 729.472940] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 729.538384] env[61936]: DEBUG nova.policy [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3598139516ee408bb6e448d3b5eeb1ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '20b0f4cf46784a2f992d4783d0bf3b1d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.714703] env[61936]: INFO nova.scheduler.client.report [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted allocations for instance a942e040-549e-4aca-8c4c-67de236e746e [ 729.814693] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae957f5-5b3f-4914-8682-f4bca16955d3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.824362] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6935ab9-a2bb-44fb-b146-cbd66c10136f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.862889] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55533652-c3ad-4192-98f8-07879c5ffe68 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.870358] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c13c20-f6bc-490d-b677-4a9a8a2502c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.883913] env[61936]: DEBUG nova.compute.provider_tree [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.972976] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 730.231458] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3b693965-6146-4c9d-b2bf-0b04b7d51e2f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "a942e040-549e-4aca-8c4c-67de236e746e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.642s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.382163] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Successfully created port: 4dc175f2-b47d-462d-bd11-13952a4d3a3e {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.387120] env[61936]: DEBUG nova.scheduler.client.report [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 730.736720] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 730.893041] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.932s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.893870] env[61936]: ERROR nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Traceback (most recent call last): [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self.driver.spawn(context, instance, image_meta, [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] vm_ref = self.build_virtual_machine(instance, [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.893870] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] for vif in network_info: [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return self._sync_wrapper(fn, *args, **kwargs) [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self.wait() [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self[:] = self._gt.wait() [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return self._exit_event.wait() [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] result = hub.switch() [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.894328] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return self.greenlet.switch() [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] result = function(*args, **kwargs) [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] return func(*args, **kwargs) [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] raise e [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] nwinfo = self.network_api.allocate_for_instance( [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] created_port_ids = self._update_ports_for_instance( [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] with excutils.save_and_reraise_exception(): [ 730.894844] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] self.force_reraise() [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] raise self.value [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] updated_port = self._update_port( [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] _ensure_no_port_binding_failure(port) [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] raise exception.PortBindingFailed(port_id=port['id']) [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] nova.exception.PortBindingFailed: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. [ 730.895484] env[61936]: ERROR nova.compute.manager [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] [ 730.895881] env[61936]: DEBUG nova.compute.utils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.895881] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.434s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.897214] env[61936]: INFO nova.compute.claims [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.900027] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Build of instance ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b was re-scheduled: Binding failed for port 3c386828-398a-43d1-8312-717874c6c43a, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 730.900478] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 730.900588] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Acquiring lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.900729] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Acquired lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.900890] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.982258] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 731.005631] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 731.005906] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.006650] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 731.006869] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.007024] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 731.007177] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 731.007384] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 731.007595] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 731.007795] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 731.007963] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 731.008146] env[61936]: DEBUG nova.virt.hardware [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 731.009579] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1baa5e7b-6d0e-4a03-a905-c412c33d63a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.018600] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7076f640-bae0-4599-ba44-901ced7f9c97 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.268039] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.431239] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.622662] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.939511] env[61936]: DEBUG nova.compute.manager [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Received event network-changed-4dc175f2-b47d-462d-bd11-13952a4d3a3e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 731.939891] env[61936]: DEBUG nova.compute.manager [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Refreshing instance network info cache due to event network-changed-4dc175f2-b47d-462d-bd11-13952a4d3a3e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 731.940139] env[61936]: DEBUG oslo_concurrency.lockutils [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] Acquiring lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.940310] env[61936]: DEBUG oslo_concurrency.lockutils [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] Acquired lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.940537] env[61936]: DEBUG nova.network.neutron [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Refreshing network info cache for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 732.024798] env[61936]: ERROR nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. [ 732.024798] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.024798] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 732.024798] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.024798] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.024798] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.024798] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.024798] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.024798] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.024798] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 732.024798] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.024798] env[61936]: ERROR nova.compute.manager raise self.value [ 732.024798] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.024798] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.024798] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.024798] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.025314] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.025314] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.025314] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. [ 732.025314] env[61936]: ERROR nova.compute.manager [ 732.025431] env[61936]: Traceback (most recent call last): [ 732.025467] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.025467] env[61936]: listener.cb(fileno) [ 732.025467] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.025467] env[61936]: result = function(*args, **kwargs) [ 732.025467] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.025467] env[61936]: return func(*args, **kwargs) [ 732.025467] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 732.025467] env[61936]: raise e [ 732.025467] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 732.025467] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 732.025467] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.025467] env[61936]: created_port_ids = self._update_ports_for_instance( [ 732.025467] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.025467] env[61936]: with excutils.save_and_reraise_exception(): [ 732.025854] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.025854] env[61936]: self.force_reraise() [ 732.025854] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.025854] env[61936]: raise self.value [ 732.025854] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.025854] env[61936]: updated_port = self._update_port( [ 732.025854] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.025854] env[61936]: _ensure_no_port_binding_failure(port) [ 732.025854] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.025854] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.025854] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. [ 732.025854] env[61936]: Removing descriptor: 19 [ 732.026396] env[61936]: ERROR nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Traceback (most recent call last): [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] yield resources [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self.driver.spawn(context, instance, image_meta, [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] vm_ref = self.build_virtual_machine(instance, [ 732.026396] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] for vif in network_info: [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return self._sync_wrapper(fn, *args, **kwargs) [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self.wait() [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self[:] = self._gt.wait() [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return self._exit_event.wait() [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.026715] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] result = hub.switch() [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return self.greenlet.switch() [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] result = function(*args, **kwargs) [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return func(*args, **kwargs) [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] raise e [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] nwinfo = self.network_api.allocate_for_instance( [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] created_port_ids = self._update_ports_for_instance( [ 732.027194] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] with excutils.save_and_reraise_exception(): [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self.force_reraise() [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] raise self.value [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] updated_port = self._update_port( [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] _ensure_no_port_binding_failure(port) [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] raise exception.PortBindingFailed(port_id=port['id']) [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] nova.exception.PortBindingFailed: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. [ 732.027604] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] [ 732.028773] env[61936]: INFO nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Terminating instance [ 732.125724] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Releasing lock "refresh_cache-ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.125868] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 732.126063] env[61936]: DEBUG nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 732.126231] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.145062] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.233304] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b03c34a-c8eb-40e6-8af7-46ad50b69e1e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.241460] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b332fc57-c7c6-4d25-b32d-7b225eaabb08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.278320] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf9d894-0a57-4366-b343-48391dfac610 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.285874] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa173491-fa96-4eb5-8a65-46d978b7de3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.299702] env[61936]: DEBUG nova.compute.provider_tree [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.462825] env[61936]: DEBUG nova.network.neutron [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.536462] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Acquiring lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.579206] env[61936]: DEBUG nova.network.neutron [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.609182] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.609525] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.647754] env[61936]: DEBUG nova.network.neutron [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.803287] env[61936]: DEBUG nova.scheduler.client.report [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.081880] env[61936]: DEBUG oslo_concurrency.lockutils [req-c59776f3-3d51-4251-85e8-b96f7a56447a req-8cf5a557-2120-4c2a-8421-5ab066ed5548 service nova] Releasing lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.081880] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Acquired lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.081880] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 733.150577] env[61936]: INFO nova.compute.manager [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] [instance: ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b] Took 1.02 seconds to deallocate network for instance. [ 733.312159] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.312159] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 733.314389] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.635s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.316572] env[61936]: INFO nova.compute.claims [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.601138] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.749725] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.821457] env[61936]: DEBUG nova.compute.utils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.831018] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 733.831018] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.952323] env[61936]: DEBUG nova.policy [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '335f4520b78f404ca67166af9f676ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca412a547c0845bb95ecf0a9f14effb1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.081773] env[61936]: DEBUG nova.compute.manager [req-2a962505-c1ef-48f1-b18f-750c3f207713 req-f65a7beb-a02c-442e-b77e-b1d0789d0328 service nova] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Received event network-vif-deleted-4dc175f2-b47d-462d-bd11-13952a4d3a3e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 734.187302] env[61936]: INFO nova.scheduler.client.report [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Deleted allocations for instance ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b [ 734.255146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Releasing lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.255633] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 734.255859] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 734.256197] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a7ff9be-296a-49fe-a593-e0b900d02c95 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.266115] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dcc65e3-e21d-471f-a972-0006a05ee2b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.296041] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9d805bf8-f2cd-4108-9007-2de18e1bf283 could not be found. [ 734.296212] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.296397] env[61936]: INFO nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Took 0.04 seconds to destroy the instance on the hypervisor. [ 734.296642] env[61936]: DEBUG oslo.service.loopingcall [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.296876] env[61936]: DEBUG nova.compute.manager [-] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 734.296979] env[61936]: DEBUG nova.network.neutron [-] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 734.318600] env[61936]: DEBUG nova.network.neutron [-] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.329655] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 734.407382] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Successfully created port: e509f45c-0eaa-4764-b585-cabe7fd86411 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.667780] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.668165] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.703680] env[61936]: DEBUG oslo_concurrency.lockutils [None req-18c6298b-5a85-4fa1-b45f-58f85901a44b tempest-InstanceActionsNegativeTestJSON-1719241418 tempest-InstanceActionsNegativeTestJSON-1719241418-project-member] Lock "ecbebef3-b0b0-45a7-b63b-a16f3ceb8c0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.881s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.735253] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6553d8c-b324-4b5e-afb1-9be9b66c5023 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.743437] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e87fd9c-9701-4619-9f50-1c03cdfb3b31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.777723] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431514b8-c81a-446f-8ec9-92b2713140a5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.784938] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad067d5-404b-488f-b9df-46539452c810 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.798194] env[61936]: DEBUG nova.compute.provider_tree [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.822593] env[61936]: DEBUG nova.network.neutron [-] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.206701] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 735.301261] env[61936]: DEBUG nova.scheduler.client.report [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 735.325442] env[61936]: INFO nova.compute.manager [-] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Took 1.03 seconds to deallocate network for instance. [ 735.333833] env[61936]: DEBUG nova.compute.claims [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.334070] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.340496] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 735.367751] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:44:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='3bd1e7f8-e262-4989-8fe9-cf58b4f6d621',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1772589397',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 735.367908] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.368072] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 735.368269] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.368392] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 735.368532] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 735.368840] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 735.369119] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 735.369466] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 735.369672] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 735.369848] env[61936]: DEBUG nova.virt.hardware [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 735.371074] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab7847a-9e58-4fdd-95e0-545af604ef8d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.380649] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7a7804-eb98-4676-848d-77e81d3298b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.733247] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.806465] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.807055] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 735.809757] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.785s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.185575] env[61936]: DEBUG nova.compute.manager [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Received event network-changed-e509f45c-0eaa-4764-b585-cabe7fd86411 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 736.185770] env[61936]: DEBUG nova.compute.manager [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Refreshing instance network info cache due to event network-changed-e509f45c-0eaa-4764-b585-cabe7fd86411. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 736.185976] env[61936]: DEBUG oslo_concurrency.lockutils [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] Acquiring lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.186131] env[61936]: DEBUG oslo_concurrency.lockutils [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] Acquired lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.186285] env[61936]: DEBUG nova.network.neutron [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Refreshing network info cache for port e509f45c-0eaa-4764-b585-cabe7fd86411 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.281494] env[61936]: ERROR nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. [ 736.281494] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 736.281494] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 736.281494] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 736.281494] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.281494] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 736.281494] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.281494] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 736.281494] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.281494] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 736.281494] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.281494] env[61936]: ERROR nova.compute.manager raise self.value [ 736.281494] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.281494] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 736.281494] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.281494] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 736.282378] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.282378] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 736.282378] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. [ 736.282378] env[61936]: ERROR nova.compute.manager [ 736.282378] env[61936]: Traceback (most recent call last): [ 736.282378] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 736.282378] env[61936]: listener.cb(fileno) [ 736.282378] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.282378] env[61936]: result = function(*args, **kwargs) [ 736.282378] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 736.282378] env[61936]: return func(*args, **kwargs) [ 736.282378] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 736.282378] env[61936]: raise e [ 736.282378] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 736.282378] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 736.282378] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.282378] env[61936]: created_port_ids = self._update_ports_for_instance( [ 736.282378] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.282378] env[61936]: with excutils.save_and_reraise_exception(): [ 736.282378] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.282378] env[61936]: self.force_reraise() [ 736.282378] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.282378] env[61936]: raise self.value [ 736.282378] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.282378] env[61936]: updated_port = self._update_port( [ 736.282378] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.282378] env[61936]: _ensure_no_port_binding_failure(port) [ 736.282378] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.282378] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 736.283559] env[61936]: nova.exception.PortBindingFailed: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. [ 736.283559] env[61936]: Removing descriptor: 19 [ 736.283559] env[61936]: ERROR nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Traceback (most recent call last): [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] yield resources [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self.driver.spawn(context, instance, image_meta, [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.283559] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] vm_ref = self.build_virtual_machine(instance, [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] for vif in network_info: [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return self._sync_wrapper(fn, *args, **kwargs) [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self.wait() [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self[:] = self._gt.wait() [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return self._exit_event.wait() [ 736.283880] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] result = hub.switch() [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return self.greenlet.switch() [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] result = function(*args, **kwargs) [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return func(*args, **kwargs) [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] raise e [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] nwinfo = self.network_api.allocate_for_instance( [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.284222] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] created_port_ids = self._update_ports_for_instance( [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] with excutils.save_and_reraise_exception(): [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self.force_reraise() [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] raise self.value [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] updated_port = self._update_port( [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] _ensure_no_port_binding_failure(port) [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.284534] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] raise exception.PortBindingFailed(port_id=port['id']) [ 736.284834] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] nova.exception.PortBindingFailed: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. [ 736.284834] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] [ 736.284834] env[61936]: INFO nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Terminating instance [ 736.314679] env[61936]: DEBUG nova.compute.utils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.323604] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 736.323604] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 736.391937] env[61936]: DEBUG nova.policy [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '288fd1d0d8e341999481fc62fe14195e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f1b64c32564b9cbb68e3799b456641', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.668424] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca0662e-caf8-4215-ae8d-ac5401be484f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.675665] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd6a2ab-fa3f-4640-b871-0c9b7338393b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.710723] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d2a865-b2fe-454c-a959-b444ca62449a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.720064] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2cd182-5cf3-4a45-8f1d-1ccead9bdc68 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.735243] env[61936]: DEBUG nova.compute.provider_tree [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.760653] env[61936]: DEBUG nova.network.neutron [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.788755] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.820512] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 736.901210] env[61936]: DEBUG nova.network.neutron [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.172337] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Successfully created port: 61f2f573-8fa4-47f8-a662-016e2df644ee {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.243504] env[61936]: DEBUG nova.scheduler.client.report [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 737.404657] env[61936]: DEBUG oslo_concurrency.lockutils [req-5f5c2397-9fba-4af7-915f-1636e1186541 req-3ccbec3a-cc44-40e7-9d46-537b581b3fae service nova] Releasing lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.405150] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquired lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.405271] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.753048] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.753602] env[61936]: ERROR nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Traceback (most recent call last): [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self.driver.spawn(context, instance, image_meta, [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] vm_ref = self.build_virtual_machine(instance, [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.753602] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] for vif in network_info: [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return self._sync_wrapper(fn, *args, **kwargs) [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self.wait() [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self[:] = self._gt.wait() [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return self._exit_event.wait() [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] result = hub.switch() [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.753956] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return self.greenlet.switch() [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] result = function(*args, **kwargs) [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] return func(*args, **kwargs) [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] raise e [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] nwinfo = self.network_api.allocate_for_instance( [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] created_port_ids = self._update_ports_for_instance( [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] with excutils.save_and_reraise_exception(): [ 737.754394] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] self.force_reraise() [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] raise self.value [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] updated_port = self._update_port( [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] _ensure_no_port_binding_failure(port) [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] raise exception.PortBindingFailed(port_id=port['id']) [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] nova.exception.PortBindingFailed: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. [ 737.754819] env[61936]: ERROR nova.compute.manager [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] [ 737.755142] env[61936]: DEBUG nova.compute.utils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.756095] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.652s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.758123] env[61936]: INFO nova.compute.claims [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.760828] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Build of instance 70e751e6-4109-41fd-a80f-82d22c34ea64 was re-scheduled: Binding failed for port 2c1e7e34-6223-4293-989c-3cb2f28ff8f8, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 737.761284] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 737.761507] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Acquiring lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.761648] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Acquired lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.761820] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.833244] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 737.866041] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 737.866285] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.866436] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 737.866608] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.866786] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 737.866878] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 737.867185] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 737.867369] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 737.867468] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 737.867648] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 737.867859] env[61936]: DEBUG nova.virt.hardware [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 737.868697] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f2aa42-fb0f-4d4b-9718-4cb396182db7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.877453] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4375719d-d6ae-48b4-8cf5-e18da23f02f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.941884] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.012526] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.216597] env[61936]: DEBUG nova.compute.manager [req-d9b3a93e-71c0-4ce9-9f45-29c413f96b27 req-b5210db5-fd89-43f6-a24c-a3d692442b0b service nova] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Received event network-vif-deleted-e509f45c-0eaa-4764-b585-cabe7fd86411 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 738.283451] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.484482] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.516669] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Releasing lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.516669] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 738.516669] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 738.516669] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2387b478-9fce-4fa0-afa3-5560e82970d0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.524692] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f8adf6-5fc8-4bb1-a464-3130c224cf1c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.549124] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 66e52362-2f67-4e5a-823a-eed7a87d1a1d could not be found. [ 738.549365] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.549547] env[61936]: INFO nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 738.549795] env[61936]: DEBUG oslo.service.loopingcall [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.550021] env[61936]: DEBUG nova.compute.manager [-] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 738.550104] env[61936]: DEBUG nova.network.neutron [-] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.573187] env[61936]: DEBUG nova.network.neutron [-] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.764609] env[61936]: ERROR nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. [ 738.764609] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.764609] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.764609] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.764609] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.764609] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.764609] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.764609] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.764609] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.764609] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 738.764609] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.764609] env[61936]: ERROR nova.compute.manager raise self.value [ 738.764609] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.764609] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.764609] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.764609] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.765025] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.765025] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.765025] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. [ 738.765025] env[61936]: ERROR nova.compute.manager [ 738.765025] env[61936]: Traceback (most recent call last): [ 738.765025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.765025] env[61936]: listener.cb(fileno) [ 738.765025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.765025] env[61936]: result = function(*args, **kwargs) [ 738.765025] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.765025] env[61936]: return func(*args, **kwargs) [ 738.765025] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 738.765025] env[61936]: raise e [ 738.765025] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.765025] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 738.765025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.765025] env[61936]: created_port_ids = self._update_ports_for_instance( [ 738.765025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.765025] env[61936]: with excutils.save_and_reraise_exception(): [ 738.765025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.765025] env[61936]: self.force_reraise() [ 738.765025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.765025] env[61936]: raise self.value [ 738.765025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.765025] env[61936]: updated_port = self._update_port( [ 738.765025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.765025] env[61936]: _ensure_no_port_binding_failure(port) [ 738.765025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.765025] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.765705] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. [ 738.765705] env[61936]: Removing descriptor: 19 [ 738.765705] env[61936]: ERROR nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Traceback (most recent call last): [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] yield resources [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self.driver.spawn(context, instance, image_meta, [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.765705] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] vm_ref = self.build_virtual_machine(instance, [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] for vif in network_info: [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return self._sync_wrapper(fn, *args, **kwargs) [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self.wait() [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self[:] = self._gt.wait() [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return self._exit_event.wait() [ 738.766036] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] result = hub.switch() [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return self.greenlet.switch() [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] result = function(*args, **kwargs) [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return func(*args, **kwargs) [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] raise e [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] nwinfo = self.network_api.allocate_for_instance( [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.766348] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] created_port_ids = self._update_ports_for_instance( [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] with excutils.save_and_reraise_exception(): [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self.force_reraise() [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] raise self.value [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] updated_port = self._update_port( [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] _ensure_no_port_binding_failure(port) [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.766732] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] raise exception.PortBindingFailed(port_id=port['id']) [ 738.767068] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] nova.exception.PortBindingFailed: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. [ 738.767068] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] [ 738.767068] env[61936]: INFO nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Terminating instance [ 738.986377] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Releasing lock "refresh_cache-70e751e6-4109-41fd-a80f-82d22c34ea64" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.986600] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 738.988066] env[61936]: DEBUG nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 738.988066] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.031200] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.065246] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4f88f0-827a-4680-8c50-3f7c7f67afe9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.072765] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5a255d-ce95-4b2c-a92e-2b3db427422e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.076181] env[61936]: DEBUG nova.network.neutron [-] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.109172] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219def8d-ad2e-461a-910c-6b7954dcf614 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.116846] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af096584-6800-45d9-a447-97cd7af5e48a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.131356] env[61936]: DEBUG nova.compute.provider_tree [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.274570] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.275840] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.275840] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.535782] env[61936]: DEBUG nova.network.neutron [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.578796] env[61936]: INFO nova.compute.manager [-] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Took 1.03 seconds to deallocate network for instance. [ 739.581645] env[61936]: DEBUG nova.compute.claims [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 739.581818] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.634576] env[61936]: DEBUG nova.scheduler.client.report [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.794202] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.007143] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.037397] env[61936]: INFO nova.compute.manager [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] [instance: 70e751e6-4109-41fd-a80f-82d22c34ea64] Took 1.05 seconds to deallocate network for instance. [ 740.139305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.139768] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 740.142253] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.851s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.297117] env[61936]: DEBUG nova.compute.manager [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Received event network-changed-61f2f573-8fa4-47f8-a662-016e2df644ee {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 740.297117] env[61936]: DEBUG nova.compute.manager [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Refreshing instance network info cache due to event network-changed-61f2f573-8fa4-47f8-a662-016e2df644ee. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 740.297471] env[61936]: DEBUG oslo_concurrency.lockutils [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] Acquiring lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.509769] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.510524] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 740.511200] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.511200] env[61936]: DEBUG oslo_concurrency.lockutils [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] Acquired lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.511373] env[61936]: DEBUG nova.network.neutron [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Refreshing network info cache for port 61f2f573-8fa4-47f8-a662-016e2df644ee {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.512295] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a5c3a91-82b4-43b9-8762-b7697921d6ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.527047] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92eebbe6-e1a4-4f39-a5d9-8678e9a38747 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.553962] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a2bb2a06-0fb8-422c-bd16-8f369e397dd2 could not be found. [ 740.554238] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.554366] env[61936]: INFO nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 740.554581] env[61936]: DEBUG oslo.service.loopingcall [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.554837] env[61936]: DEBUG nova.compute.manager [-] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 740.555641] env[61936]: DEBUG nova.network.neutron [-] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.571163] env[61936]: DEBUG nova.network.neutron [-] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.648059] env[61936]: DEBUG nova.compute.utils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.648059] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 740.648059] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.709362] env[61936]: DEBUG nova.policy [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b98c1bbe1044663830b18c4ce9e23d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1b7d01ffac74c6fa2306b4839041fdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.049988] env[61936]: DEBUG nova.network.neutron [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.075090] env[61936]: DEBUG nova.network.neutron [-] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.081583] env[61936]: INFO nova.scheduler.client.report [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Deleted allocations for instance 70e751e6-4109-41fd-a80f-82d22c34ea64 [ 741.091342] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe148cff-3deb-4f4a-824d-00d7055e051b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.103860] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3451ab9-7565-413e-8a39-ecea2f4c86c3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.139774] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdbfc63-2091-46a6-a5bb-922e3685e83b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.148267] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecbdbbe-1b8e-47af-bf84-aaaabce3b6d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.161905] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 741.166831] env[61936]: DEBUG nova.compute.provider_tree [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.194491] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Successfully created port: 5e06644e-9ca4-43d1-9c6c-130e7391dc7c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.226471] env[61936]: DEBUG nova.network.neutron [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.578019] env[61936]: INFO nova.compute.manager [-] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Took 1.02 seconds to deallocate network for instance. [ 741.580180] env[61936]: DEBUG nova.compute.claims [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.580368] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.598078] env[61936]: DEBUG oslo_concurrency.lockutils [None req-671260b2-9f38-4434-825f-1385c054f900 tempest-ServerActionsTestOtherB-1837475219 tempest-ServerActionsTestOtherB-1837475219-project-member] Lock "70e751e6-4109-41fd-a80f-82d22c34ea64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.975s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.672146] env[61936]: DEBUG nova.scheduler.client.report [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 741.729104] env[61936]: DEBUG oslo_concurrency.lockutils [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] Releasing lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.729414] env[61936]: DEBUG nova.compute.manager [req-52a37a96-8e8d-4291-95df-7a76dd8dc040 req-cb485a23-cf6e-4bae-aa82-43f1c626958e service nova] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Received event network-vif-deleted-61f2f573-8fa4-47f8-a662-016e2df644ee {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 742.100927] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 742.179434] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 742.183501] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.041s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.184236] env[61936]: ERROR nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Traceback (most recent call last): [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self.driver.spawn(context, instance, image_meta, [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] vm_ref = self.build_virtual_machine(instance, [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.184236] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] for vif in network_info: [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return self._sync_wrapper(fn, *args, **kwargs) [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self.wait() [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self[:] = self._gt.wait() [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return self._exit_event.wait() [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] result = hub.switch() [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.185761] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return self.greenlet.switch() [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] result = function(*args, **kwargs) [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] return func(*args, **kwargs) [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] raise e [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] nwinfo = self.network_api.allocate_for_instance( [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] created_port_ids = self._update_ports_for_instance( [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] with excutils.save_and_reraise_exception(): [ 742.186551] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] self.force_reraise() [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] raise self.value [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] updated_port = self._update_port( [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] _ensure_no_port_binding_failure(port) [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] raise exception.PortBindingFailed(port_id=port['id']) [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] nova.exception.PortBindingFailed: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. [ 742.187135] env[61936]: ERROR nova.compute.manager [instance: 64e78c61-85f4-4914-8fab-3ced64335275] [ 742.187512] env[61936]: DEBUG nova.compute.utils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.188016] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.714s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.189601] env[61936]: INFO nova.compute.claims [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.192541] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Build of instance 64e78c61-85f4-4914-8fab-3ced64335275 was re-scheduled: Binding failed for port f140be6f-84da-41d7-be72-63d83dafab5b, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 742.192790] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 742.193036] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Acquiring lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.193182] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Acquired lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.193372] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.206404] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 742.206536] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.206777] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 742.207095] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.207297] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 742.207464] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 742.207702] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 742.207850] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 742.208033] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 742.208198] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 742.208366] env[61936]: DEBUG nova.virt.hardware [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 742.209264] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207793be-74e0-49ae-ad6e-93cf042c97be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.218491] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5696173b-50fd-41a8-81f5-3ea637baad55 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.560360] env[61936]: DEBUG nova.compute.manager [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Received event network-changed-5e06644e-9ca4-43d1-9c6c-130e7391dc7c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 742.560360] env[61936]: DEBUG nova.compute.manager [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Refreshing instance network info cache due to event network-changed-5e06644e-9ca4-43d1-9c6c-130e7391dc7c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 742.560360] env[61936]: DEBUG oslo_concurrency.lockutils [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] Acquiring lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.560360] env[61936]: DEBUG oslo_concurrency.lockutils [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] Acquired lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.560360] env[61936]: DEBUG nova.network.neutron [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Refreshing network info cache for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 742.598235] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "bdea3742-5789-4e97-b139-33cfee134843" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.600721] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "bdea3742-5789-4e97-b139-33cfee134843" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.624733] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.633943] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "77a59650-d28b-4862-bb35-6910e6f7d6c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.638623] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.658224] env[61936]: ERROR nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. [ 742.658224] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 742.658224] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.658224] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 742.658224] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.658224] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 742.658224] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.658224] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 742.658224] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.658224] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 742.658224] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.658224] env[61936]: ERROR nova.compute.manager raise self.value [ 742.658224] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.658224] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 742.658224] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.658224] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 742.658765] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.658765] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 742.658765] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. [ 742.658765] env[61936]: ERROR nova.compute.manager [ 742.658765] env[61936]: Traceback (most recent call last): [ 742.658765] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 742.658765] env[61936]: listener.cb(fileno) [ 742.658765] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.658765] env[61936]: result = function(*args, **kwargs) [ 742.658765] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.658765] env[61936]: return func(*args, **kwargs) [ 742.658765] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.658765] env[61936]: raise e [ 742.658765] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.658765] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 742.658765] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.658765] env[61936]: created_port_ids = self._update_ports_for_instance( [ 742.658765] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.658765] env[61936]: with excutils.save_and_reraise_exception(): [ 742.658765] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.658765] env[61936]: self.force_reraise() [ 742.658765] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.658765] env[61936]: raise self.value [ 742.658765] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.658765] env[61936]: updated_port = self._update_port( [ 742.658765] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.658765] env[61936]: _ensure_no_port_binding_failure(port) [ 742.658765] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.658765] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 742.661066] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. [ 742.661066] env[61936]: Removing descriptor: 19 [ 742.661066] env[61936]: ERROR nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Traceback (most recent call last): [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] yield resources [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self.driver.spawn(context, instance, image_meta, [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.661066] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] vm_ref = self.build_virtual_machine(instance, [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] for vif in network_info: [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return self._sync_wrapper(fn, *args, **kwargs) [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self.wait() [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self[:] = self._gt.wait() [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return self._exit_event.wait() [ 742.661512] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] result = hub.switch() [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return self.greenlet.switch() [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] result = function(*args, **kwargs) [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return func(*args, **kwargs) [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] raise e [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] nwinfo = self.network_api.allocate_for_instance( [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.661904] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] created_port_ids = self._update_ports_for_instance( [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] with excutils.save_and_reraise_exception(): [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self.force_reraise() [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] raise self.value [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] updated_port = self._update_port( [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] _ensure_no_port_binding_failure(port) [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.662256] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] raise exception.PortBindingFailed(port_id=port['id']) [ 742.663375] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] nova.exception.PortBindingFailed: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. [ 742.663375] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] [ 742.663375] env[61936]: INFO nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Terminating instance [ 742.667348] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.667573] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.712969] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.793545] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.079777] env[61936]: DEBUG nova.network.neutron [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.165379] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.184133] env[61936]: DEBUG nova.network.neutron [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.296547] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Releasing lock "refresh_cache-64e78c61-85f4-4914-8fab-3ced64335275" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.296785] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 743.296961] env[61936]: DEBUG nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 743.297141] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.317882] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.619246] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc2feb8-98cb-4baa-b57d-df88495d0843 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.627372] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f507a1ec-f292-4ff5-a1c1-199a50589bfa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.659936] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a65fb6-cf26-4a44-98a7-a0e602eaed2a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.665514] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b926b3-cda9-40b2-9b87-8c21697c653b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.679764] env[61936]: DEBUG nova.compute.provider_tree [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.685756] env[61936]: DEBUG oslo_concurrency.lockutils [req-bb78be12-94e7-4d0d-9e0d-48577ac09043 req-066338e1-644d-4065-b0eb-cc11e6705993 service nova] Releasing lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.686135] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.686316] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.821194] env[61936]: DEBUG nova.network.neutron [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.183701] env[61936]: DEBUG nova.scheduler.client.report [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 744.212487] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.311853] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.323865] env[61936]: INFO nova.compute.manager [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] [instance: 64e78c61-85f4-4914-8fab-3ced64335275] Took 1.03 seconds to deallocate network for instance. [ 744.582891] env[61936]: DEBUG nova.compute.manager [req-d11d1c07-1569-422f-9dea-301a8340ee88 req-832e615b-3435-45ca-bb49-0532ea0ea035 service nova] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Received event network-vif-deleted-5e06644e-9ca4-43d1-9c6c-130e7391dc7c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 744.691470] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.691986] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 744.694728] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.393s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.694872] env[61936]: DEBUG nova.objects.instance [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 744.814395] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.814684] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 744.814883] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.815346] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cecb0750-4caf-486f-a9a9-475dbe2194c5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.827426] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b89950-6a2f-4795-bbdf-acd090fcca31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.853149] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bdf4c76e-2441-4f76-bf74-2718c57c13e2 could not be found. [ 744.853398] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.853581] env[61936]: INFO nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 744.853823] env[61936]: DEBUG oslo.service.loopingcall [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.854194] env[61936]: DEBUG nova.compute.manager [-] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 744.854194] env[61936]: DEBUG nova.network.neutron [-] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.868766] env[61936]: DEBUG nova.network.neutron [-] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.200054] env[61936]: DEBUG nova.compute.utils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.204177] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 745.204381] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.271825] env[61936]: DEBUG nova.policy [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3382c0703a24837a7a62574bfc6ed37', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f03f2cf1d4084fdca2664ccaac0a7671', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.360376] env[61936]: INFO nova.scheduler.client.report [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Deleted allocations for instance 64e78c61-85f4-4914-8fab-3ced64335275 [ 745.370748] env[61936]: DEBUG nova.network.neutron [-] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.556537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.556537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.662315] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Successfully created port: 4b5a2d2e-2584-4820-8b86-c486d1f842ba {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.705097] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 745.711048] env[61936]: DEBUG oslo_concurrency.lockutils [None req-39eeedf1-a70b-4e4b-aef4-4aef26f6b8cc tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.711048] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.256s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.711048] env[61936]: DEBUG nova.objects.instance [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 745.869698] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1abcb07c-3f88-40c0-85a6-044cdd010f98 tempest-AttachInterfacesV270Test-1674150742 tempest-AttachInterfacesV270Test-1674150742-project-member] Lock "64e78c61-85f4-4914-8fab-3ced64335275" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.154s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.875121] env[61936]: INFO nova.compute.manager [-] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Took 1.02 seconds to deallocate network for instance. [ 745.875841] env[61936]: DEBUG nova.compute.claims [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 745.876173] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.150860] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Successfully created port: b87a0e7c-d722-46a7-b6ca-3732ed25d67c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.371707] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 746.498290] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Successfully created port: 7476e0a6-b5b7-49c2-8832-3af2bce948c8 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.720086] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 746.723745] env[61936]: DEBUG oslo_concurrency.lockutils [None req-95639d0e-496d-4126-9ed5-8b70201290b1 tempest-ServersAdmin275Test-738653653 tempest-ServersAdmin275Test-738653653-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.725728] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.049s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.726205] env[61936]: DEBUG nova.objects.instance [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lazy-loading 'resources' on Instance uuid 1e0dade4-ea54-457d-8398-54ba810a83c2 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 746.754844] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 746.755863] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.755863] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 746.756976] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.756976] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 746.756976] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 746.756976] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 746.757421] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 746.757722] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 746.760017] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 746.760017] env[61936]: DEBUG nova.virt.hardware [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 746.760017] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bea136-0945-4595-a0b2-832d0b26b795 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.768750] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb9db88-cedf-49b3-88ee-1644d855c7ed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.901226] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.527049] env[61936]: DEBUG nova.compute.manager [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Received event network-changed-4b5a2d2e-2584-4820-8b86-c486d1f842ba {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 747.527289] env[61936]: DEBUG nova.compute.manager [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Refreshing instance network info cache due to event network-changed-4b5a2d2e-2584-4820-8b86-c486d1f842ba. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 747.527541] env[61936]: DEBUG oslo_concurrency.lockutils [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] Acquiring lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.527744] env[61936]: DEBUG oslo_concurrency.lockutils [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] Acquired lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.527998] env[61936]: DEBUG nova.network.neutron [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Refreshing network info cache for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.641451] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da8c1ab-d305-4e49-95ab-c6ee45bd7062 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.649348] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f60674d-48f6-4909-a03b-b3973e394dd9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.683592] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea3d932-3e5f-4640-8104-1793b94ec36f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.695763] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbaac0a-a77b-4b8e-9c62-bdffd785fd06 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.711368] env[61936]: DEBUG nova.compute.provider_tree [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.732139] env[61936]: ERROR nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. [ 747.732139] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 747.732139] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 747.732139] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 747.732139] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.732139] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 747.732139] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.732139] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 747.732139] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.732139] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 747.732139] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.732139] env[61936]: ERROR nova.compute.manager raise self.value [ 747.732139] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.732139] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 747.732139] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.732139] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 747.732752] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.732752] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 747.732752] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. [ 747.732752] env[61936]: ERROR nova.compute.manager [ 747.732752] env[61936]: Traceback (most recent call last): [ 747.732752] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 747.732752] env[61936]: listener.cb(fileno) [ 747.732752] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.732752] env[61936]: result = function(*args, **kwargs) [ 747.732752] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.732752] env[61936]: return func(*args, **kwargs) [ 747.732752] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 747.732752] env[61936]: raise e [ 747.732752] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 747.732752] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 747.732752] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.732752] env[61936]: created_port_ids = self._update_ports_for_instance( [ 747.732752] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.732752] env[61936]: with excutils.save_and_reraise_exception(): [ 747.732752] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.732752] env[61936]: self.force_reraise() [ 747.732752] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.732752] env[61936]: raise self.value [ 747.732752] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.732752] env[61936]: updated_port = self._update_port( [ 747.732752] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.732752] env[61936]: _ensure_no_port_binding_failure(port) [ 747.732752] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.732752] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 747.734370] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. [ 747.734370] env[61936]: Removing descriptor: 19 [ 747.734370] env[61936]: ERROR nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Traceback (most recent call last): [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] yield resources [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self.driver.spawn(context, instance, image_meta, [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.734370] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] vm_ref = self.build_virtual_machine(instance, [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] for vif in network_info: [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return self._sync_wrapper(fn, *args, **kwargs) [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self.wait() [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self[:] = self._gt.wait() [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return self._exit_event.wait() [ 747.734898] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] result = hub.switch() [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return self.greenlet.switch() [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] result = function(*args, **kwargs) [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return func(*args, **kwargs) [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] raise e [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] nwinfo = self.network_api.allocate_for_instance( [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.735444] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] created_port_ids = self._update_ports_for_instance( [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] with excutils.save_and_reraise_exception(): [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self.force_reraise() [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] raise self.value [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] updated_port = self._update_port( [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] _ensure_no_port_binding_failure(port) [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.736045] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] raise exception.PortBindingFailed(port_id=port['id']) [ 747.736471] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] nova.exception.PortBindingFailed: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. [ 747.736471] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] [ 747.736471] env[61936]: INFO nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Terminating instance [ 748.087644] env[61936]: DEBUG nova.network.neutron [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.189264] env[61936]: DEBUG nova.network.neutron [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.213871] env[61936]: DEBUG nova.scheduler.client.report [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 748.238415] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.692294] env[61936]: DEBUG oslo_concurrency.lockutils [req-8096dd04-6b0c-4d31-acd4-44ed7c1f77b9 req-99bd5b03-7359-4e18-9a4c-ab4f285bfc9a service nova] Releasing lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.692816] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquired lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.692953] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.718011] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.720314] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.384s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.721854] env[61936]: INFO nova.compute.claims [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.743840] env[61936]: INFO nova.scheduler.client.report [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Deleted allocations for instance 1e0dade4-ea54-457d-8398-54ba810a83c2 [ 749.231835] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.256957] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2225ce55-cfed-466f-a523-b7d6b98f76f3 tempest-ServersAdmin275Test-611109554 tempest-ServersAdmin275Test-611109554-project-member] Lock "1e0dade4-ea54-457d-8398-54ba810a83c2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.767s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.326940] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.591676] env[61936]: DEBUG nova.compute.manager [req-a041c6c5-6f3f-4dad-97e1-105dececef01 req-442a2f8f-f8e6-45fd-a667-ecab6c7f0e6b service nova] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Received event network-vif-deleted-4b5a2d2e-2584-4820-8b86-c486d1f842ba {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 749.830020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Releasing lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.830957] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 749.831172] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 749.835329] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-137b69cc-bdff-4744-8863-b79b376b6468 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.851663] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94443496-c6c8-403e-9783-620aea881b68 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.878145] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6c3575f-9351-4957-a200-4e4d0557b55b could not be found. [ 749.878381] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.878559] env[61936]: INFO nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 749.878860] env[61936]: DEBUG oslo.service.loopingcall [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.883499] env[61936]: DEBUG nova.compute.manager [-] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 749.883600] env[61936]: DEBUG nova.network.neutron [-] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.928167] env[61936]: DEBUG nova.network.neutron [-] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.122404] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561b416d-ffec-4417-8c1a-6b64d16e7163 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.131979] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1051267-fd1e-4280-93ad-4c6af2944fd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.169016] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ff679c-8df3-4923-925d-743dbfe8687e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.177545] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d350bd1c-bf4d-476d-b27d-73bd5772a631 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.192978] env[61936]: DEBUG nova.compute.provider_tree [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.696092] env[61936]: DEBUG nova.scheduler.client.report [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 751.137029] env[61936]: DEBUG nova.network.neutron [-] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.203401] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.203982] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 751.207062] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.939s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.208543] env[61936]: INFO nova.compute.claims [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.330314] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "e5e3e7af-8104-4f7a-ae18-81e03a932608" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.330549] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.641378] env[61936]: INFO nova.compute.manager [-] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Took 1.76 seconds to deallocate network for instance. [ 751.643792] env[61936]: DEBUG nova.compute.claims [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 751.643969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.712765] env[61936]: DEBUG nova.compute.utils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.717214] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 751.717214] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 751.765477] env[61936]: DEBUG nova.policy [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3bcf6acdbc494bd2a78f95073a4b4873', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc5491d9385446eba3492f693ba9dd3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.219651] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 752.226597] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Successfully created port: 38a28341-39c4-4389-adf5-9e57a0233a88 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.616053] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3479a4-25df-4e67-a0d2-f4f0e369d91d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.626060] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d172db89-03e3-460f-acd5-8894538891d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.660109] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db73e94-0f2b-45ec-b4bb-9ec038659ca5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.667402] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f225e5a-adc4-4ca5-9bfc-6f6b3f5e3ba9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.681354] env[61936]: DEBUG nova.compute.provider_tree [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.185878] env[61936]: DEBUG nova.scheduler.client.report [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 753.237020] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 753.262177] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 753.262998] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.263211] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 753.264317] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.264317] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 753.264317] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 753.264538] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 753.265451] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 753.265451] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 753.265451] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 753.265451] env[61936]: DEBUG nova.virt.hardware [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 753.267086] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4097f07b-65a0-4a57-bfab-17ddc5d55d46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.276594] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1080c3d-8ddc-4756-a916-8b09526be93c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.479306] env[61936]: DEBUG nova.compute.manager [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Received event network-changed-38a28341-39c4-4389-adf5-9e57a0233a88 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 753.479502] env[61936]: DEBUG nova.compute.manager [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Refreshing instance network info cache due to event network-changed-38a28341-39c4-4389-adf5-9e57a0233a88. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 753.479708] env[61936]: DEBUG oslo_concurrency.lockutils [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] Acquiring lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.479851] env[61936]: DEBUG oslo_concurrency.lockutils [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] Acquired lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.480018] env[61936]: DEBUG nova.network.neutron [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Refreshing network info cache for port 38a28341-39c4-4389-adf5-9e57a0233a88 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.638232] env[61936]: ERROR nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. [ 753.638232] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.638232] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.638232] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.638232] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.638232] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.638232] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.638232] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.638232] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.638232] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 753.638232] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.638232] env[61936]: ERROR nova.compute.manager raise self.value [ 753.638232] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.638232] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.638232] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.638232] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.638748] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.638748] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.638748] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. [ 753.638748] env[61936]: ERROR nova.compute.manager [ 753.638748] env[61936]: Traceback (most recent call last): [ 753.638748] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.638748] env[61936]: listener.cb(fileno) [ 753.638748] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.638748] env[61936]: result = function(*args, **kwargs) [ 753.638748] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.638748] env[61936]: return func(*args, **kwargs) [ 753.638748] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 753.638748] env[61936]: raise e [ 753.638748] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.638748] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 753.638748] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.638748] env[61936]: created_port_ids = self._update_ports_for_instance( [ 753.638748] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.638748] env[61936]: with excutils.save_and_reraise_exception(): [ 753.638748] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.638748] env[61936]: self.force_reraise() [ 753.638748] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.638748] env[61936]: raise self.value [ 753.638748] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.638748] env[61936]: updated_port = self._update_port( [ 753.638748] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.638748] env[61936]: _ensure_no_port_binding_failure(port) [ 753.638748] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.638748] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.639665] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. [ 753.639665] env[61936]: Removing descriptor: 19 [ 753.639665] env[61936]: ERROR nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] Traceback (most recent call last): [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] yield resources [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self.driver.spawn(context, instance, image_meta, [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.639665] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] vm_ref = self.build_virtual_machine(instance, [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] for vif in network_info: [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return self._sync_wrapper(fn, *args, **kwargs) [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self.wait() [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self[:] = self._gt.wait() [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return self._exit_event.wait() [ 753.640009] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] result = hub.switch() [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return self.greenlet.switch() [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] result = function(*args, **kwargs) [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return func(*args, **kwargs) [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] raise e [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] nwinfo = self.network_api.allocate_for_instance( [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.640373] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] created_port_ids = self._update_ports_for_instance( [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] with excutils.save_and_reraise_exception(): [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self.force_reraise() [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] raise self.value [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] updated_port = self._update_port( [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] _ensure_no_port_binding_failure(port) [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.640715] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] raise exception.PortBindingFailed(port_id=port['id']) [ 753.642061] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] nova.exception.PortBindingFailed: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. [ 753.642061] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] [ 753.642061] env[61936]: INFO nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Terminating instance [ 753.690920] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.691497] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 753.694534] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.360s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.001880] env[61936]: DEBUG nova.network.neutron [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.085533] env[61936]: DEBUG nova.network.neutron [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.143378] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Acquiring lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.200473] env[61936]: DEBUG nova.compute.utils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.207080] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 754.207080] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 754.246543] env[61936]: DEBUG nova.policy [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91be2be86c4c4613aa5f8ce7a911567f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cd5d8086e444b84843a5bc61360b702', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 754.588296] env[61936]: DEBUG oslo_concurrency.lockutils [req-2a1480a4-8bf2-4ba6-a0d3-2f2ef259dd6f req-8ba3d938-b59d-4bfe-b638-b06a2c0d7419 service nova] Releasing lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.588862] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Acquired lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.589074] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.596514] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33966b5f-9407-425c-bc48-cce186740485 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.604904] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe85133-2d81-4f06-bb3d-bee8ec976d03 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.635744] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1146578-179d-4924-9c6c-105e4176055c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.645371] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a58e597-2884-493d-9752-079526059e00 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.659776] env[61936]: DEBUG nova.compute.provider_tree [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.707110] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 754.840534] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Successfully created port: c2d6747c-d7d4-47a0-83cf-7742ec62c406 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.123750] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.163495] env[61936]: DEBUG nova.scheduler.client.report [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 755.334867] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.671493] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.672135] env[61936]: ERROR nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Traceback (most recent call last): [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self.driver.spawn(context, instance, image_meta, [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] vm_ref = self.build_virtual_machine(instance, [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.672135] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] for vif in network_info: [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return self._sync_wrapper(fn, *args, **kwargs) [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self.wait() [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self[:] = self._gt.wait() [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return self._exit_event.wait() [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] result = hub.switch() [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 755.672547] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return self.greenlet.switch() [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] result = function(*args, **kwargs) [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] return func(*args, **kwargs) [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] raise e [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] nwinfo = self.network_api.allocate_for_instance( [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] created_port_ids = self._update_ports_for_instance( [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] with excutils.save_and_reraise_exception(): [ 755.672905] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] self.force_reraise() [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] raise self.value [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] updated_port = self._update_port( [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] _ensure_no_port_binding_failure(port) [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] raise exception.PortBindingFailed(port_id=port['id']) [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] nova.exception.PortBindingFailed: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. [ 755.673499] env[61936]: ERROR nova.compute.manager [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] [ 755.673976] env[61936]: DEBUG nova.compute.utils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 755.674865] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.942s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.677024] env[61936]: INFO nova.compute.claims [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.681163] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Build of instance 9d805bf8-f2cd-4108-9007-2de18e1bf283 was re-scheduled: Binding failed for port 4dc175f2-b47d-462d-bd11-13952a4d3a3e, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 755.681779] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 755.682091] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Acquiring lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.682299] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Acquired lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.682520] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 755.719778] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 755.743458] env[61936]: DEBUG nova.compute.manager [req-a83befb8-5350-4c1d-81b0-01cd92da220b req-4e56f87a-2aa0-442b-bbd2-71d1a1284b08 service nova] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Received event network-vif-deleted-38a28341-39c4-4389-adf5-9e57a0233a88 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 755.751813] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 755.752048] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.753216] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 755.753216] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.753216] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 755.753216] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 755.753216] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 755.753436] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 755.753436] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 755.753436] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 755.753613] env[61936]: DEBUG nova.virt.hardware [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 755.754747] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a02df3-ca91-4201-9abc-abd9cc90ff14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.767113] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5a3eaf-baa8-4a6d-a15c-9d1b85ff6a64 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.840375] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Releasing lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.840716] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 755.840914] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.841237] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a057ca71-739a-492b-abab-2423076d0261 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.855037] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea0ccb9-abd6-422a-b433-84ff043b90af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.878027] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7599598-492e-4d42-8417-960df12b2cd7 could not be found. [ 755.878284] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.878462] env[61936]: INFO nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 755.878788] env[61936]: DEBUG oslo.service.loopingcall [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.879063] env[61936]: DEBUG nova.compute.manager [-] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 755.879164] env[61936]: DEBUG nova.network.neutron [-] [instance: d7599598-492e-4d42-8417-960df12b2cd7] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.906550] env[61936]: DEBUG nova.network.neutron [-] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.211379] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.234608] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "f2844173-4114-4158-99ad-ac651efbb607" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.235307] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "f2844173-4114-4158-99ad-ac651efbb607" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.368606] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.409210] env[61936]: DEBUG nova.network.neutron [-] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.451397] env[61936]: ERROR nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. [ 756.451397] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 756.451397] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 756.451397] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 756.451397] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.451397] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 756.451397] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.451397] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 756.451397] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.451397] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 756.451397] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.451397] env[61936]: ERROR nova.compute.manager raise self.value [ 756.451397] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.451397] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 756.451397] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.451397] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 756.452025] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.452025] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 756.452025] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. [ 756.452025] env[61936]: ERROR nova.compute.manager [ 756.452025] env[61936]: Traceback (most recent call last): [ 756.452025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 756.452025] env[61936]: listener.cb(fileno) [ 756.452025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.452025] env[61936]: result = function(*args, **kwargs) [ 756.452025] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 756.452025] env[61936]: return func(*args, **kwargs) [ 756.452025] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 756.452025] env[61936]: raise e [ 756.452025] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 756.452025] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 756.452025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.452025] env[61936]: created_port_ids = self._update_ports_for_instance( [ 756.452025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.452025] env[61936]: with excutils.save_and_reraise_exception(): [ 756.452025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.452025] env[61936]: self.force_reraise() [ 756.452025] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.452025] env[61936]: raise self.value [ 756.452025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.452025] env[61936]: updated_port = self._update_port( [ 756.452025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.452025] env[61936]: _ensure_no_port_binding_failure(port) [ 756.452025] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.452025] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 756.453204] env[61936]: nova.exception.PortBindingFailed: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. [ 756.453204] env[61936]: Removing descriptor: 19 [ 756.453204] env[61936]: ERROR nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Traceback (most recent call last): [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] yield resources [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self.driver.spawn(context, instance, image_meta, [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.453204] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] vm_ref = self.build_virtual_machine(instance, [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] for vif in network_info: [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return self._sync_wrapper(fn, *args, **kwargs) [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self.wait() [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self[:] = self._gt.wait() [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return self._exit_event.wait() [ 756.453651] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] result = hub.switch() [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return self.greenlet.switch() [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] result = function(*args, **kwargs) [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return func(*args, **kwargs) [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] raise e [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] nwinfo = self.network_api.allocate_for_instance( [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.454033] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] created_port_ids = self._update_ports_for_instance( [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] with excutils.save_and_reraise_exception(): [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self.force_reraise() [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] raise self.value [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] updated_port = self._update_port( [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] _ensure_no_port_binding_failure(port) [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.454484] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] raise exception.PortBindingFailed(port_id=port['id']) [ 756.454871] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] nova.exception.PortBindingFailed: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. [ 756.454871] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] [ 756.454871] env[61936]: INFO nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Terminating instance [ 756.871426] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Releasing lock "refresh_cache-9d805bf8-f2cd-4108-9007-2de18e1bf283" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.871426] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 756.871426] env[61936]: DEBUG nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 756.871426] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.901401] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.911784] env[61936]: INFO nova.compute.manager [-] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Took 1.03 seconds to deallocate network for instance. [ 756.914197] env[61936]: DEBUG nova.compute.claims [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 756.914197] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.956497] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Acquiring lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.956656] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Acquired lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.956822] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.005723] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e2310c-880c-43e2-a076-21a3400730e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.013240] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb8cd72-a6b0-435b-91f1-92764a631c31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.042100] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86872fe6-81e0-4494-bee7-2de669c1280d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.048870] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81eb0587-3889-4370-bb97-b9cdf66241f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.061355] env[61936]: DEBUG nova.compute.provider_tree [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.404047] env[61936]: DEBUG nova.network.neutron [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.473294] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.554034] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.564299] env[61936]: DEBUG nova.scheduler.client.report [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 757.783215] env[61936]: DEBUG nova.compute.manager [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Received event network-changed-c2d6747c-d7d4-47a0-83cf-7742ec62c406 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 757.783455] env[61936]: DEBUG nova.compute.manager [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Refreshing instance network info cache due to event network-changed-c2d6747c-d7d4-47a0-83cf-7742ec62c406. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 757.783670] env[61936]: DEBUG oslo_concurrency.lockutils [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] Acquiring lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.906872] env[61936]: INFO nova.compute.manager [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] [instance: 9d805bf8-f2cd-4108-9007-2de18e1bf283] Took 1.04 seconds to deallocate network for instance. [ 758.056474] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Releasing lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.057014] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 758.057297] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.057464] env[61936]: DEBUG oslo_concurrency.lockutils [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] Acquired lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.057604] env[61936]: DEBUG nova.network.neutron [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Refreshing network info cache for port c2d6747c-d7d4-47a0-83cf-7742ec62c406 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 758.058705] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dab92dc9-93ac-4ea0-a810-dc3f21c081fa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.068374] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3118a4-2d7d-4925-835a-2d760a9a53e5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.079060] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.079531] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 758.082015] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.500s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.096266] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 201d06f9-d277-4d8e-84ee-84e1053c1137 could not be found. [ 758.096478] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.096651] env[61936]: INFO nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Took 0.04 seconds to destroy the instance on the hypervisor. [ 758.096884] env[61936]: DEBUG oslo.service.loopingcall [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.097112] env[61936]: DEBUG nova.compute.manager [-] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 758.097204] env[61936]: DEBUG nova.network.neutron [-] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.131997] env[61936]: DEBUG nova.network.neutron [-] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.576938] env[61936]: DEBUG nova.network.neutron [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.587131] env[61936]: DEBUG nova.compute.utils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.591833] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 758.591999] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 758.636114] env[61936]: DEBUG nova.network.neutron [-] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.671688] env[61936]: DEBUG nova.policy [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e387455720b437a95ee2e15609155ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ac7f88f349d4e2eae5c062f4d3d8cd9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.747133] env[61936]: DEBUG nova.network.neutron [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.976022] env[61936]: INFO nova.scheduler.client.report [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Deleted allocations for instance 9d805bf8-f2cd-4108-9007-2de18e1bf283 [ 759.047754] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8c4765-f04d-44bb-867f-ce9165c6f609 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.055601] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b533bb34-2f63-48f6-afe4-c1b973f50554 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.089626] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3df3185-da73-4b09-a737-13399f5bfcdb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.094651] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 759.101027] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7033031-c1e9-431f-ad13-75283a92cd38 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.115369] env[61936]: DEBUG nova.compute.provider_tree [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.139041] env[61936]: INFO nova.compute.manager [-] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Took 1.04 seconds to deallocate network for instance. [ 759.141357] env[61936]: DEBUG nova.compute.claims [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 759.141577] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.187011] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Successfully created port: c0032bcf-b39a-4df0-8207-41222dcc060c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.256077] env[61936]: DEBUG oslo_concurrency.lockutils [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] Releasing lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.256370] env[61936]: DEBUG nova.compute.manager [req-2fd16934-a776-4653-8d92-1fc50d30caae req-a568c1ea-8cc1-44fe-9fbf-4e3b2e05aa9b service nova] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Received event network-vif-deleted-c2d6747c-d7d4-47a0-83cf-7742ec62c406 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 759.487517] env[61936]: DEBUG oslo_concurrency.lockutils [None req-77ba180e-b761-4fc1-91d0-e9fe62c5170b tempest-ServersTestManualDisk-293697012 tempest-ServersTestManualDisk-293697012-project-member] Lock "9d805bf8-f2cd-4108-9007-2de18e1bf283" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.062s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.620871] env[61936]: DEBUG nova.scheduler.client.report [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 759.990140] env[61936]: DEBUG nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 760.106358] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 760.124370] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.124974] env[61936]: ERROR nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Traceback (most recent call last): [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self.driver.spawn(context, instance, image_meta, [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] vm_ref = self.build_virtual_machine(instance, [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.124974] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] for vif in network_info: [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return self._sync_wrapper(fn, *args, **kwargs) [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self.wait() [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self[:] = self._gt.wait() [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return self._exit_event.wait() [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] result = hub.switch() [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.125313] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return self.greenlet.switch() [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] result = function(*args, **kwargs) [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] return func(*args, **kwargs) [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] raise e [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] nwinfo = self.network_api.allocate_for_instance( [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] created_port_ids = self._update_ports_for_instance( [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] with excutils.save_and_reraise_exception(): [ 760.125800] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] self.force_reraise() [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] raise self.value [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] updated_port = self._update_port( [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] _ensure_no_port_binding_failure(port) [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] raise exception.PortBindingFailed(port_id=port['id']) [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] nova.exception.PortBindingFailed: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. [ 760.126161] env[61936]: ERROR nova.compute.manager [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] [ 760.126478] env[61936]: DEBUG nova.compute.utils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.130776] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.550s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.134191] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Build of instance 66e52362-2f67-4e5a-823a-eed7a87d1a1d was re-scheduled: Binding failed for port e509f45c-0eaa-4764-b585-cabe7fd86411, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 760.135060] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 760.135681] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquiring lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.138230] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Acquired lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.138410] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.145809] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 760.146070] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.146170] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 760.146339] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.146480] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 760.146651] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 760.146860] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 760.147720] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 760.147991] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 760.148219] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 760.148401] env[61936]: DEBUG nova.virt.hardware [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 760.153020] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40150ae8-ce41-478a-b323-b020eab498f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.159808] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775902ab-2c8b-42b8-963d-9417f9561598 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.167582] env[61936]: DEBUG nova.compute.manager [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Received event network-changed-c0032bcf-b39a-4df0-8207-41222dcc060c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 760.167880] env[61936]: DEBUG nova.compute.manager [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Refreshing instance network info cache due to event network-changed-c0032bcf-b39a-4df0-8207-41222dcc060c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 760.168141] env[61936]: DEBUG oslo_concurrency.lockutils [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] Acquiring lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.168325] env[61936]: DEBUG oslo_concurrency.lockutils [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] Acquired lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.168508] env[61936]: DEBUG nova.network.neutron [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Refreshing network info cache for port c0032bcf-b39a-4df0-8207-41222dcc060c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.368747] env[61936]: ERROR nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. [ 760.368747] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 760.368747] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 760.368747] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 760.368747] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.368747] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 760.368747] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.368747] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 760.368747] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.368747] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 760.368747] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.368747] env[61936]: ERROR nova.compute.manager raise self.value [ 760.368747] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.368747] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 760.368747] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.368747] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 760.369283] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.369283] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 760.369283] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. [ 760.369283] env[61936]: ERROR nova.compute.manager [ 760.369283] env[61936]: Traceback (most recent call last): [ 760.369283] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 760.369283] env[61936]: listener.cb(fileno) [ 760.369283] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.369283] env[61936]: result = function(*args, **kwargs) [ 760.369283] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.369283] env[61936]: return func(*args, **kwargs) [ 760.369283] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 760.369283] env[61936]: raise e [ 760.369283] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 760.369283] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 760.369283] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.369283] env[61936]: created_port_ids = self._update_ports_for_instance( [ 760.369283] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.369283] env[61936]: with excutils.save_and_reraise_exception(): [ 760.369283] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.369283] env[61936]: self.force_reraise() [ 760.369283] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.369283] env[61936]: raise self.value [ 760.369283] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.369283] env[61936]: updated_port = self._update_port( [ 760.369283] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.369283] env[61936]: _ensure_no_port_binding_failure(port) [ 760.369283] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.369283] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 760.370264] env[61936]: nova.exception.PortBindingFailed: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. [ 760.370264] env[61936]: Removing descriptor: 15 [ 760.370264] env[61936]: ERROR nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Traceback (most recent call last): [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] yield resources [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self.driver.spawn(context, instance, image_meta, [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self._vmops.spawn(context, instance, image_meta, injected_files, [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 760.370264] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] vm_ref = self.build_virtual_machine(instance, [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] vif_infos = vmwarevif.get_vif_info(self._session, [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] for vif in network_info: [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return self._sync_wrapper(fn, *args, **kwargs) [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self.wait() [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self[:] = self._gt.wait() [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return self._exit_event.wait() [ 760.370685] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] result = hub.switch() [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return self.greenlet.switch() [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] result = function(*args, **kwargs) [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return func(*args, **kwargs) [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] raise e [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] nwinfo = self.network_api.allocate_for_instance( [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 760.371123] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] created_port_ids = self._update_ports_for_instance( [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] with excutils.save_and_reraise_exception(): [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self.force_reraise() [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] raise self.value [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] updated_port = self._update_port( [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] _ensure_no_port_binding_failure(port) [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 760.371558] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] raise exception.PortBindingFailed(port_id=port['id']) [ 760.371943] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] nova.exception.PortBindingFailed: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. [ 760.371943] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] [ 760.371943] env[61936]: INFO nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Terminating instance [ 760.521053] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.659883] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.697797] env[61936]: DEBUG nova.network.neutron [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.763572] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.827139] env[61936]: DEBUG nova.network.neutron [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.874191] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Acquiring lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.994164] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741cb2b9-becf-482d-a03d-c16795d040a5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.002634] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b1a5d8-1d6c-4b70-9a3f-bf8d68ba0bb5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.035155] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31cb4ac-b23b-4539-9fa9-07a5d32147fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.042639] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b6acbb-a2c2-4de3-9d06-929f85a28155 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.056019] env[61936]: DEBUG nova.compute.provider_tree [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.267090] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Releasing lock "refresh_cache-66e52362-2f67-4e5a-823a-eed7a87d1a1d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.267413] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 761.267593] env[61936]: DEBUG nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 761.267941] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.287848] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.329767] env[61936]: DEBUG oslo_concurrency.lockutils [req-0a94775b-61c4-4a82-8b6f-e780cf0160ea req-1ed2554f-31ea-4609-abbc-350381f07ed4 service nova] Releasing lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.330265] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Acquired lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.330458] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.559814] env[61936]: DEBUG nova.scheduler.client.report [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.792171] env[61936]: DEBUG nova.network.neutron [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.860507] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.964831] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.066712] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.067413] env[61936]: ERROR nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Traceback (most recent call last): [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self.driver.spawn(context, instance, image_meta, [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] vm_ref = self.build_virtual_machine(instance, [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.067413] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] for vif in network_info: [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return self._sync_wrapper(fn, *args, **kwargs) [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self.wait() [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self[:] = self._gt.wait() [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return self._exit_event.wait() [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] result = hub.switch() [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.067821] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return self.greenlet.switch() [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] result = function(*args, **kwargs) [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] return func(*args, **kwargs) [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] raise e [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] nwinfo = self.network_api.allocate_for_instance( [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] created_port_ids = self._update_ports_for_instance( [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] with excutils.save_and_reraise_exception(): [ 762.068496] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] self.force_reraise() [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] raise self.value [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] updated_port = self._update_port( [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] _ensure_no_port_binding_failure(port) [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] raise exception.PortBindingFailed(port_id=port['id']) [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] nova.exception.PortBindingFailed: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. [ 762.068853] env[61936]: ERROR nova.compute.manager [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] [ 762.070169] env[61936]: DEBUG nova.compute.utils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 762.072107] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.447s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.073606] env[61936]: INFO nova.compute.claims [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.076265] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Build of instance a2bb2a06-0fb8-422c-bd16-8f369e397dd2 was re-scheduled: Binding failed for port 61f2f573-8fa4-47f8-a662-016e2df644ee, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 762.076752] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 762.077021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.077224] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.077416] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.187867] env[61936]: DEBUG nova.compute.manager [req-1394086b-4a0b-45d4-bee9-f92e7fa4cb86 req-5c3ec186-45d2-4bd8-8a95-872416652e5e service nova] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Received event network-vif-deleted-c0032bcf-b39a-4df0-8207-41222dcc060c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 762.294964] env[61936]: INFO nova.compute.manager [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] [instance: 66e52362-2f67-4e5a-823a-eed7a87d1a1d] Took 1.03 seconds to deallocate network for instance. [ 762.468189] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Releasing lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.468855] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 762.469101] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.469402] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f45f607-a30d-4c25-bdbc-815631b5ef6a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.478271] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3dffde-015c-419f-a125-a2fdd400b986 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.499928] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 03ab71e9-64b7-4534-a693-473f3903c511 could not be found. [ 762.500160] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.500362] env[61936]: INFO nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Took 0.03 seconds to destroy the instance on the hypervisor. [ 762.500600] env[61936]: DEBUG oslo.service.loopingcall [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.500875] env[61936]: DEBUG nova.compute.manager [-] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 762.500973] env[61936]: DEBUG nova.network.neutron [-] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.514792] env[61936]: DEBUG nova.network.neutron [-] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.602014] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.705463] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.019711] env[61936]: DEBUG nova.network.neutron [-] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.206534] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-a2bb2a06-0fb8-422c-bd16-8f369e397dd2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.206766] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 763.206978] env[61936]: DEBUG nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 763.207193] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.227378] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.321216] env[61936]: INFO nova.scheduler.client.report [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Deleted allocations for instance 66e52362-2f67-4e5a-823a-eed7a87d1a1d [ 763.430305] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c168763-438a-4151-afff-75cb7ab6d92a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.439054] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13171246-1831-4a5c-a0c8-d2e7cd9069e6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.468551] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19631563-3538-4b6e-9666-8668596853ba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.476019] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac93a8e2-1110-4448-ac71-b459c4d9dc5f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.489555] env[61936]: DEBUG nova.compute.provider_tree [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.523075] env[61936]: INFO nova.compute.manager [-] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Took 1.02 seconds to deallocate network for instance. [ 763.525593] env[61936]: DEBUG nova.compute.claims [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.525772] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.730243] env[61936]: DEBUG nova.network.neutron [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.830715] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9005940e-1250-41f4-8c47-f4750df1e27c tempest-MigrationsAdminTest-1243380832 tempest-MigrationsAdminTest-1243380832-project-member] Lock "66e52362-2f67-4e5a-823a-eed7a87d1a1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.772s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.992530] env[61936]: DEBUG nova.scheduler.client.report [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.233213] env[61936]: INFO nova.compute.manager [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: a2bb2a06-0fb8-422c-bd16-8f369e397dd2] Took 1.03 seconds to deallocate network for instance. [ 764.333988] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 764.498168] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.498710] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 764.501272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.625s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.864943] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.005362] env[61936]: DEBUG nova.compute.utils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.009938] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 765.010396] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.247959] env[61936]: DEBUG nova.policy [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0b8a725defa427cbb829d6eb286dce5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dcf17a5d80e048928c2c0a9fabffb14c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.265709] env[61936]: INFO nova.scheduler.client.report [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted allocations for instance a2bb2a06-0fb8-422c-bd16-8f369e397dd2 [ 765.348301] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15033bd2-9e98-4a65-b742-41b4b7b4a838 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.356468] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4470a02e-a690-44f6-b095-9b6c6f701c02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.387549] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e56b9a8-3c32-46da-a5d1-fa9121807761 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.396561] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36954054-9de4-4670-9e69-17f54c62d490 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.413749] env[61936]: DEBUG nova.compute.provider_tree [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.514626] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 765.685411] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Successfully created port: 5c4ef559-9891-4f7f-8608-19a8fbd74016 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.776366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-aa3317e5-f768-4dd6-934a-2215ce7a680b tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "a2bb2a06-0fb8-422c-bd16-8f369e397dd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.653s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.920180] env[61936]: DEBUG nova.scheduler.client.report [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 766.280102] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 766.427724] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.428805] env[61936]: ERROR nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Traceback (most recent call last): [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self.driver.spawn(context, instance, image_meta, [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] vm_ref = self.build_virtual_machine(instance, [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.428805] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] for vif in network_info: [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return self._sync_wrapper(fn, *args, **kwargs) [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self.wait() [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self[:] = self._gt.wait() [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return self._exit_event.wait() [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] result = hub.switch() [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 766.429269] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return self.greenlet.switch() [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] result = function(*args, **kwargs) [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] return func(*args, **kwargs) [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] raise e [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] nwinfo = self.network_api.allocate_for_instance( [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] created_port_ids = self._update_ports_for_instance( [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] with excutils.save_and_reraise_exception(): [ 766.429804] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] self.force_reraise() [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] raise self.value [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] updated_port = self._update_port( [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] _ensure_no_port_binding_failure(port) [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] raise exception.PortBindingFailed(port_id=port['id']) [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] nova.exception.PortBindingFailed: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. [ 766.430219] env[61936]: ERROR nova.compute.manager [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] [ 766.430580] env[61936]: DEBUG nova.compute.utils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.432016] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Build of instance bdf4c76e-2441-4f76-bf74-2718c57c13e2 was re-scheduled: Binding failed for port 5e06644e-9ca4-43d1-9c6c-130e7391dc7c, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 766.432913] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 766.433660] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.433833] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.433998] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.434999] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.535s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.438411] env[61936]: INFO nova.compute.claims [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.471321] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.523688] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 766.545251] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 766.545567] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.545772] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.546336] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.546500] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.546697] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 766.546843] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 766.546998] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 766.547189] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 766.547454] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 766.547586] env[61936]: DEBUG nova.virt.hardware [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 766.548794] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e2385d-ad28-4d17-bcc5-6f548f749642 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.558334] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe34ba5-647d-4244-b12c-29f61704bec7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.677476] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.804765] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.993410] env[61936]: DEBUG nova.compute.manager [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Received event network-changed-5c4ef559-9891-4f7f-8608-19a8fbd74016 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 766.993891] env[61936]: DEBUG nova.compute.manager [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Refreshing instance network info cache due to event network-changed-5c4ef559-9891-4f7f-8608-19a8fbd74016. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 766.994170] env[61936]: DEBUG oslo_concurrency.lockutils [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] Acquiring lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.994367] env[61936]: DEBUG oslo_concurrency.lockutils [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] Acquired lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.994801] env[61936]: DEBUG nova.network.neutron [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Refreshing network info cache for port 5c4ef559-9891-4f7f-8608-19a8fbd74016 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 767.163384] env[61936]: ERROR nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. [ 767.163384] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 767.163384] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 767.163384] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 767.163384] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.163384] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 767.163384] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.163384] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 767.163384] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.163384] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 767.163384] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.163384] env[61936]: ERROR nova.compute.manager raise self.value [ 767.163384] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.163384] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 767.163384] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.163384] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 767.163870] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.163870] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 767.163870] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. [ 767.163870] env[61936]: ERROR nova.compute.manager [ 767.163870] env[61936]: Traceback (most recent call last): [ 767.163870] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 767.163870] env[61936]: listener.cb(fileno) [ 767.163870] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.163870] env[61936]: result = function(*args, **kwargs) [ 767.163870] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.163870] env[61936]: return func(*args, **kwargs) [ 767.163870] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 767.163870] env[61936]: raise e [ 767.163870] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 767.163870] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 767.163870] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.163870] env[61936]: created_port_ids = self._update_ports_for_instance( [ 767.163870] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.163870] env[61936]: with excutils.save_and_reraise_exception(): [ 767.163870] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.163870] env[61936]: self.force_reraise() [ 767.163870] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.163870] env[61936]: raise self.value [ 767.163870] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.163870] env[61936]: updated_port = self._update_port( [ 767.163870] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.163870] env[61936]: _ensure_no_port_binding_failure(port) [ 767.163870] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.163870] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 767.164764] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. [ 767.164764] env[61936]: Removing descriptor: 19 [ 767.164764] env[61936]: ERROR nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Traceback (most recent call last): [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] yield resources [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self.driver.spawn(context, instance, image_meta, [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.164764] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] vm_ref = self.build_virtual_machine(instance, [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] for vif in network_info: [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return self._sync_wrapper(fn, *args, **kwargs) [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self.wait() [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self[:] = self._gt.wait() [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return self._exit_event.wait() [ 767.165186] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] result = hub.switch() [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return self.greenlet.switch() [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] result = function(*args, **kwargs) [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return func(*args, **kwargs) [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] raise e [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] nwinfo = self.network_api.allocate_for_instance( [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.165544] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] created_port_ids = self._update_ports_for_instance( [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] with excutils.save_and_reraise_exception(): [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self.force_reraise() [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] raise self.value [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] updated_port = self._update_port( [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] _ensure_no_port_binding_failure(port) [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.165936] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] raise exception.PortBindingFailed(port_id=port['id']) [ 767.166273] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] nova.exception.PortBindingFailed: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. [ 767.166273] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] [ 767.166273] env[61936]: INFO nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Terminating instance [ 767.181120] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "refresh_cache-bdf4c76e-2441-4f76-bf74-2718c57c13e2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.181264] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 767.181349] env[61936]: DEBUG nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 767.181522] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 767.200344] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.536851] env[61936]: DEBUG nova.network.neutron [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.599720] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "e800f5e9-deb0-4592-aded-3768d238e311" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.600020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "e800f5e9-deb0-4592-aded-3768d238e311" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.657886] env[61936]: DEBUG nova.network.neutron [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.669769] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Acquiring lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.705327] env[61936]: DEBUG nova.network.neutron [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.857804] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddce1b3-d80e-49d5-91fe-421957730496 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.868181] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca853666-ca79-4302-aabb-beb71ca88c3b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.904550] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1667d186-1ab1-45bf-90a2-1f775933b454 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.912670] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0914cdb9-9203-4b67-83c7-9dc281fddf6d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.928557] env[61936]: DEBUG nova.compute.provider_tree [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.160105] env[61936]: DEBUG oslo_concurrency.lockutils [req-a4153950-b40c-4a50-a716-427a69433b87 req-9f6fb4fd-2d3e-4d2a-9c7d-8e650147db09 service nova] Releasing lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.162796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Acquired lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.162796] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.208548] env[61936]: INFO nova.compute.manager [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: bdf4c76e-2441-4f76-bf74-2718c57c13e2] Took 1.03 seconds to deallocate network for instance. [ 768.432092] env[61936]: DEBUG nova.scheduler.client.report [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 768.695791] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.826707] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.938321] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.938796] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 768.942134] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.298s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.072030] env[61936]: DEBUG nova.compute.manager [req-7807ca28-97ee-48f4-9b49-c0c6aaff69c2 req-b50a9c44-8c7c-4712-a677-9f6df0c8f57d service nova] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Received event network-vif-deleted-5c4ef559-9891-4f7f-8608-19a8fbd74016 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 769.248283] env[61936]: INFO nova.scheduler.client.report [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted allocations for instance bdf4c76e-2441-4f76-bf74-2718c57c13e2 [ 769.331537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Releasing lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.331848] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 769.332075] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 769.332691] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4fc9cd5-6dbc-45dd-a245-a17ce50470fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.342505] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e871a3c6-7e53-4650-be53-bbacabde65f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.366269] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca could not be found. [ 769.366550] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.366735] env[61936]: INFO nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Took 0.03 seconds to destroy the instance on the hypervisor. [ 769.366977] env[61936]: DEBUG oslo.service.loopingcall [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.367216] env[61936]: DEBUG nova.compute.manager [-] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 769.367310] env[61936]: DEBUG nova.network.neutron [-] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.387636] env[61936]: DEBUG nova.network.neutron [-] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.448411] env[61936]: DEBUG nova.compute.utils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.449501] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 769.456021] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 769.518198] env[61936]: DEBUG nova.policy [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d7a6b4aa1e6416191fe2e366ac153b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5f09457a25d49bca5a37817419a43c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.757294] env[61936]: DEBUG oslo_concurrency.lockutils [None req-63333795-24c8-4181-85b2-d8a104812c0d tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "bdf4c76e-2441-4f76-bf74-2718c57c13e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.110s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.823371] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba08eaf-414b-42b6-8d44-be03c4229820 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.832565] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae48c4d1-a6e9-432f-a038-3f695d39d13d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.869701] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f165d86-5612-425f-bbb1-2e9bfbebe9e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.877405] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6011f85d-4ac4-4209-b511-0e00e7ab88b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.894559] env[61936]: DEBUG nova.network.neutron [-] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.896021] env[61936]: DEBUG nova.compute.provider_tree [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.918457] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.918637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.950311] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 770.014645] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Successfully created port: 6fc5c96c-3d2d-46da-9375-74c69c0d59d7 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.260822] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 770.398322] env[61936]: INFO nova.compute.manager [-] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Took 1.03 seconds to deallocate network for instance. [ 770.399330] env[61936]: DEBUG nova.scheduler.client.report [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 770.406020] env[61936]: DEBUG nova.compute.claims [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 770.406020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.455747] env[61936]: INFO nova.virt.block_device [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Booting with volume f3f72d22-314e-410e-98fa-0223f070d3c5 at /dev/sda [ 770.518019] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cab25de7-a8cd-46ab-b68a-ce465ca521ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.527224] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5485fe67-d8f8-4bc3-902f-d6febd9aa790 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.547703] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52b2e873-4c33-4fb6-a142-b965361f04e9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.555090] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb909f1f-bf46-4709-b9fa-7423512f4ee8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.576011] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61a4c1b-222b-4012-bde4-d6b6555b299f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.581922] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9683ac99-3220-4a5a-8fce-ab03752688e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.595358] env[61936]: DEBUG nova.virt.block_device [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Updating existing volume attachment record: f89f5e9e-538c-4445-adf3-4161c0a36c0e {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 770.785365] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.901324] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.901525] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.905466] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.964s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.906105] env[61936]: ERROR nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Traceback (most recent call last): [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self.driver.spawn(context, instance, image_meta, [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] vm_ref = self.build_virtual_machine(instance, [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.906105] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] for vif in network_info: [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return self._sync_wrapper(fn, *args, **kwargs) [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self.wait() [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self[:] = self._gt.wait() [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return self._exit_event.wait() [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] result = hub.switch() [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 770.906482] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return self.greenlet.switch() [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] result = function(*args, **kwargs) [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] return func(*args, **kwargs) [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] raise e [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] nwinfo = self.network_api.allocate_for_instance( [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] created_port_ids = self._update_ports_for_instance( [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] with excutils.save_and_reraise_exception(): [ 770.906984] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] self.force_reraise() [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] raise self.value [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] updated_port = self._update_port( [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] _ensure_no_port_binding_failure(port) [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] raise exception.PortBindingFailed(port_id=port['id']) [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] nova.exception.PortBindingFailed: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. [ 770.907426] env[61936]: ERROR nova.compute.manager [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] [ 770.907806] env[61936]: DEBUG nova.compute.utils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 770.909690] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.994s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.914913] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Build of instance f6c3575f-9351-4957-a200-4e4d0557b55b was re-scheduled: Binding failed for port 4b5a2d2e-2584-4820-8b86-c486d1f842ba, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 770.915348] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 770.915956] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.915956] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquired lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.915956] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.263256] env[61936]: DEBUG nova.compute.manager [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Received event network-changed-6fc5c96c-3d2d-46da-9375-74c69c0d59d7 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 771.263256] env[61936]: DEBUG nova.compute.manager [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Refreshing instance network info cache due to event network-changed-6fc5c96c-3d2d-46da-9375-74c69c0d59d7. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 771.263256] env[61936]: DEBUG oslo_concurrency.lockutils [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] Acquiring lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.263256] env[61936]: DEBUG oslo_concurrency.lockutils [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] Acquired lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.263256] env[61936]: DEBUG nova.network.neutron [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Refreshing network info cache for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 771.433516] env[61936]: ERROR nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. [ 771.433516] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 771.433516] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 771.433516] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 771.433516] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.433516] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 771.433516] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.433516] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 771.433516] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.433516] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 771.433516] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.433516] env[61936]: ERROR nova.compute.manager raise self.value [ 771.433516] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.433516] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 771.433516] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.433516] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 771.434094] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.434094] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 771.434094] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. [ 771.434094] env[61936]: ERROR nova.compute.manager [ 771.434094] env[61936]: Traceback (most recent call last): [ 771.434094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 771.434094] env[61936]: listener.cb(fileno) [ 771.434094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.434094] env[61936]: result = function(*args, **kwargs) [ 771.434094] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 771.434094] env[61936]: return func(*args, **kwargs) [ 771.434094] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 771.434094] env[61936]: raise e [ 771.434094] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 771.434094] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 771.434094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.434094] env[61936]: created_port_ids = self._update_ports_for_instance( [ 771.434094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.434094] env[61936]: with excutils.save_and_reraise_exception(): [ 771.434094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.434094] env[61936]: self.force_reraise() [ 771.434094] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.434094] env[61936]: raise self.value [ 771.434094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.434094] env[61936]: updated_port = self._update_port( [ 771.434094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.434094] env[61936]: _ensure_no_port_binding_failure(port) [ 771.434094] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.434094] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 771.434985] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. [ 771.434985] env[61936]: Removing descriptor: 19 [ 771.442814] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.530603] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.746351] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373d2e23-ce3d-4bad-957e-447c953a70eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.754257] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e26d4f7-bd73-48c1-94a8-767398e5edb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.786692] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b45fce6-090f-47c3-8672-094f1054822f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.794228] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369fa876-719c-44ae-8597-c17e57604100 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.807974] env[61936]: DEBUG nova.compute.provider_tree [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.814052] env[61936]: DEBUG nova.network.neutron [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.921934] env[61936]: DEBUG nova.network.neutron [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.033706] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Releasing lock "refresh_cache-f6c3575f-9351-4957-a200-4e4d0557b55b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.033993] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 772.034192] env[61936]: DEBUG nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 772.034353] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.051529] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.311121] env[61936]: DEBUG nova.scheduler.client.report [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 772.427318] env[61936]: DEBUG oslo_concurrency.lockutils [req-f18d552d-4663-44be-930e-3b640696805a req-1ea95356-c7e7-4199-82e6-3cf2be7c6b07 service nova] Releasing lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.553750] env[61936]: DEBUG nova.network.neutron [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.702689] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 772.703301] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 772.703532] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.703686] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 772.703860] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.704013] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 772.704165] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 772.704363] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 772.704514] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 772.704671] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 772.704825] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 772.704993] env[61936]: DEBUG nova.virt.hardware [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 772.705863] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5552fb-97e4-45ac-80bf-023a166dede5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.714661] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1713c433-86d3-479e-83bf-28b766d53859 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.728559] env[61936]: ERROR nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Traceback (most recent call last): [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] yield resources [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self.driver.spawn(context, instance, image_meta, [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] vm_ref = self.build_virtual_machine(instance, [ 772.728559] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] for vif in network_info: [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] return self._sync_wrapper(fn, *args, **kwargs) [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self.wait() [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self[:] = self._gt.wait() [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] return self._exit_event.wait() [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 772.728998] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] current.throw(*self._exc) [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] result = function(*args, **kwargs) [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] return func(*args, **kwargs) [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] raise e [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] nwinfo = self.network_api.allocate_for_instance( [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] created_port_ids = self._update_ports_for_instance( [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] with excutils.save_and_reraise_exception(): [ 772.729369] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self.force_reraise() [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] raise self.value [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] updated_port = self._update_port( [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] _ensure_no_port_binding_failure(port) [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] raise exception.PortBindingFailed(port_id=port['id']) [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] nova.exception.PortBindingFailed: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. [ 772.729817] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] [ 772.729817] env[61936]: INFO nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Terminating instance [ 772.816602] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.817178] env[61936]: ERROR nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] Traceback (most recent call last): [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self.driver.spawn(context, instance, image_meta, [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] vm_ref = self.build_virtual_machine(instance, [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.817178] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] for vif in network_info: [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return self._sync_wrapper(fn, *args, **kwargs) [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self.wait() [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self[:] = self._gt.wait() [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return self._exit_event.wait() [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] result = hub.switch() [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.817545] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return self.greenlet.switch() [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] result = function(*args, **kwargs) [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] return func(*args, **kwargs) [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] raise e [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] nwinfo = self.network_api.allocate_for_instance( [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] created_port_ids = self._update_ports_for_instance( [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] with excutils.save_and_reraise_exception(): [ 772.817942] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] self.force_reraise() [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] raise self.value [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] updated_port = self._update_port( [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] _ensure_no_port_binding_failure(port) [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] raise exception.PortBindingFailed(port_id=port['id']) [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] nova.exception.PortBindingFailed: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. [ 772.818402] env[61936]: ERROR nova.compute.manager [instance: d7599598-492e-4d42-8417-960df12b2cd7] [ 772.818709] env[61936]: DEBUG nova.compute.utils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 772.819077] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.677s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.821950] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Build of instance d7599598-492e-4d42-8417-960df12b2cd7 was re-scheduled: Binding failed for port 38a28341-39c4-4389-adf5-9e57a0233a88, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 772.822399] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 772.822617] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Acquiring lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.822760] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Acquired lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.822914] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.056689] env[61936]: INFO nova.compute.manager [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: f6c3575f-9351-4957-a200-4e4d0557b55b] Took 1.02 seconds to deallocate network for instance. [ 773.234026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Acquiring lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.234026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Acquired lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.234026] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.283694] env[61936]: DEBUG nova.compute.manager [req-c535cabd-743b-4561-9580-df3019094b8f req-a97aa564-8f73-4f3d-8ac7-abdb06428cde service nova] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Received event network-vif-deleted-6fc5c96c-3d2d-46da-9375-74c69c0d59d7 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 773.343705] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.430478] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.621939] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adc2306-a2e0-4332-b08b-513867f5b214 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.629794] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dfccc26-be47-42e5-835e-d48e0c2b1f98 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.659684] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0bc36e-ef6e-41a5-a823-a0f4af545b1c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.667094] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2163998-f74b-424a-bb29-68bf058be469 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.679892] env[61936]: DEBUG nova.compute.provider_tree [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.749773] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.847982] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.933451] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Releasing lock "refresh_cache-d7599598-492e-4d42-8417-960df12b2cd7" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.933762] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 773.933892] env[61936]: DEBUG nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 773.934075] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 773.955355] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.089843] env[61936]: INFO nova.scheduler.client.report [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Deleted allocations for instance f6c3575f-9351-4957-a200-4e4d0557b55b [ 774.183022] env[61936]: DEBUG nova.scheduler.client.report [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 774.351544] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Releasing lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.351544] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 774.351849] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52cefc0b-2c1d-4191-97d0-40bcfd1b5048 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.360753] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8523a85c-0140-4231-85f3-3bf301e5f25c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.382103] env[61936]: WARNING nova.virt.vmwareapi.driver [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 0b14583c-dd0f-429d-a386-2acda353465a could not be found. [ 774.382338] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.382639] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c24a32e-a9f5-49c3-b028-c0cd3f5e6bf8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.390862] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0dd2d64-1cfe-4b59-9f86-cfd79cbcdea0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.411790] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b14583c-dd0f-429d-a386-2acda353465a could not be found. [ 774.412007] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.412219] env[61936]: INFO nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Took 0.06 seconds to destroy the instance on the hypervisor. [ 774.412455] env[61936]: DEBUG oslo.service.loopingcall [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.412662] env[61936]: DEBUG nova.compute.manager [-] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 774.412754] env[61936]: DEBUG nova.network.neutron [-] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.427060] env[61936]: DEBUG nova.network.neutron [-] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.457497] env[61936]: DEBUG nova.network.neutron [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.598572] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5bc350fd-a810-4363-8bfa-c5eeae7048a8 tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "f6c3575f-9351-4957-a200-4e4d0557b55b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.412s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.688852] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.689508] env[61936]: ERROR nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Traceback (most recent call last): [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self.driver.spawn(context, instance, image_meta, [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] vm_ref = self.build_virtual_machine(instance, [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.689508] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] for vif in network_info: [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return self._sync_wrapper(fn, *args, **kwargs) [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self.wait() [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self[:] = self._gt.wait() [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return self._exit_event.wait() [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] result = hub.switch() [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 774.690128] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return self.greenlet.switch() [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] result = function(*args, **kwargs) [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] return func(*args, **kwargs) [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] raise e [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] nwinfo = self.network_api.allocate_for_instance( [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] created_port_ids = self._update_ports_for_instance( [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] with excutils.save_and_reraise_exception(): [ 774.690873] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] self.force_reraise() [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] raise self.value [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] updated_port = self._update_port( [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] _ensure_no_port_binding_failure(port) [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] raise exception.PortBindingFailed(port_id=port['id']) [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] nova.exception.PortBindingFailed: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. [ 774.691539] env[61936]: ERROR nova.compute.manager [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] [ 774.692248] env[61936]: DEBUG nova.compute.utils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 774.692248] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.171s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.693539] env[61936]: INFO nova.compute.claims [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.696077] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Build of instance 201d06f9-d277-4d8e-84ee-84e1053c1137 was re-scheduled: Binding failed for port c2d6747c-d7d4-47a0-83cf-7742ec62c406, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 774.696627] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 774.696720] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Acquiring lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.696850] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Acquired lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.697012] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.929657] env[61936]: DEBUG nova.network.neutron [-] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.960464] env[61936]: INFO nova.compute.manager [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] [instance: d7599598-492e-4d42-8417-960df12b2cd7] Took 1.03 seconds to deallocate network for instance. [ 775.101721] env[61936]: DEBUG nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 775.220879] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.406198] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.433567] env[61936]: INFO nova.compute.manager [-] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Took 1.02 seconds to deallocate network for instance. [ 775.624517] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.910792] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Releasing lock "refresh_cache-201d06f9-d277-4d8e-84ee-84e1053c1137" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.910792] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 775.910792] env[61936]: DEBUG nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 775.910792] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.926032] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.996277] env[61936]: INFO nova.scheduler.client.report [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Deleted allocations for instance d7599598-492e-4d42-8417-960df12b2cd7 [ 776.004112] env[61936]: INFO nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Took 0.57 seconds to detach 1 volumes for instance. [ 776.005939] env[61936]: DEBUG nova.compute.claims [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 776.009610] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.039569] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b031b0-d9fa-4d48-bec1-d6cd9930b564 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.048825] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6159d1-9558-4fd6-be5a-4ab246a48d6f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.081674] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2887b4b2-9d64-410c-a433-31b55ab0cd6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.089673] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0b585c-e8b5-4f36-a3cd-1adf12602f31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.103341] env[61936]: DEBUG nova.compute.provider_tree [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.433370] env[61936]: DEBUG nova.network.neutron [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.503741] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8f403f23-ffc3-4751-ad66-0fa17aa20f3e tempest-ServerRescueTestJSONUnderV235-1236930080 tempest-ServerRescueTestJSONUnderV235-1236930080-project-member] Lock "d7599598-492e-4d42-8417-960df12b2cd7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.836s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.606366] env[61936]: DEBUG nova.scheduler.client.report [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 776.940842] env[61936]: INFO nova.compute.manager [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] [instance: 201d06f9-d277-4d8e-84ee-84e1053c1137] Took 1.03 seconds to deallocate network for instance. [ 777.011314] env[61936]: DEBUG nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 777.111881] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.111975] env[61936]: DEBUG nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 777.115220] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.589s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.147155] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.148242] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.535607] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.623450] env[61936]: DEBUG nova.compute.utils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.630430] env[61936]: DEBUG nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Not allocating networking since 'none' was specified. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 777.981406] env[61936]: INFO nova.scheduler.client.report [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Deleted allocations for instance 201d06f9-d277-4d8e-84ee-84e1053c1137 [ 778.029313] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c6cfc5-8592-48ef-9908-a2f419ba2cc2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.038744] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74114481-b507-4690-b29b-625bfcf1d1db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.074358] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f43742e-16e7-46de-a6ae-9a72974af642 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.081427] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abaa7bc6-1fb6-407d-8b4f-b043f7e9490f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.094402] env[61936]: DEBUG nova.compute.provider_tree [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.132781] env[61936]: DEBUG nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 778.492200] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5e38f017-c899-4702-8762-da4fd8718203 tempest-AttachInterfacesUnderV243Test-210535510 tempest-AttachInterfacesUnderV243Test-210535510-project-member] Lock "201d06f9-d277-4d8e-84ee-84e1053c1137" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.005s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.600110] env[61936]: DEBUG nova.scheduler.client.report [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 778.994791] env[61936]: DEBUG nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 779.104765] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.990s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.105463] env[61936]: ERROR nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Traceback (most recent call last): [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self.driver.spawn(context, instance, image_meta, [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] vm_ref = self.build_virtual_machine(instance, [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.105463] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] for vif in network_info: [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return self._sync_wrapper(fn, *args, **kwargs) [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self.wait() [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self[:] = self._gt.wait() [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return self._exit_event.wait() [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] result = hub.switch() [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.105806] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return self.greenlet.switch() [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] result = function(*args, **kwargs) [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] return func(*args, **kwargs) [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] raise e [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] nwinfo = self.network_api.allocate_for_instance( [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] created_port_ids = self._update_ports_for_instance( [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] with excutils.save_and_reraise_exception(): [ 779.106224] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] self.force_reraise() [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] raise self.value [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] updated_port = self._update_port( [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] _ensure_no_port_binding_failure(port) [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] raise exception.PortBindingFailed(port_id=port['id']) [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] nova.exception.PortBindingFailed: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. [ 779.106626] env[61936]: ERROR nova.compute.manager [instance: 03ab71e9-64b7-4534-a693-473f3903c511] [ 779.106925] env[61936]: DEBUG nova.compute.utils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.107520] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.243s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.109730] env[61936]: INFO nova.compute.claims [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.112548] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Build of instance 03ab71e9-64b7-4534-a693-473f3903c511 was re-scheduled: Binding failed for port c0032bcf-b39a-4df0-8207-41222dcc060c, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 779.112548] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 779.112548] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Acquiring lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.112796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Acquired lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.112796] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.140975] env[61936]: DEBUG nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 779.174969] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 779.175232] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.175385] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 779.175561] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.175700] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 779.175862] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 779.176096] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 779.176258] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 779.176421] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 779.176574] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 779.176737] env[61936]: DEBUG nova.virt.hardware [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 779.177882] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a500bc88-a549-4c7f-8893-2146341283ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.186099] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bf4800-f422-491a-b52b-fa8ca2df7541 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.199795] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.205936] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Creating folder: Project (a8ce58298d1140a2b47c4ca6e7fab632). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.205936] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfe376a0-8979-47e6-970d-6daad09fb2c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.216930] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Created folder: Project (a8ce58298d1140a2b47c4ca6e7fab632) in parent group-v269874. [ 779.216930] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Creating folder: Instances. Parent ref: group-v269891. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.217142] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10f57577-de95-4ce7-b398-718ba758eb2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.226052] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Created folder: Instances in parent group-v269891. [ 779.226296] env[61936]: DEBUG oslo.service.loopingcall [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.226475] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 779.226673] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51ab0fdf-43b5-42f5-843a-77c2f2e381d3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.242835] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.242835] env[61936]: value = "task-1252772" [ 779.242835] env[61936]: _type = "Task" [ 779.242835] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.250377] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252772, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.518571] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.639664] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.754191] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252772, 'name': CreateVM_Task, 'duration_secs': 0.248255} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.754317] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.754745] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.754907] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.755279] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.758231] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88021806-c970-4b1d-b53c-4eb1072a454c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.764966] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 779.764966] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527507f5-b8b4-9c45-92db-e5664b0e537d" [ 779.764966] env[61936]: _type = "Task" [ 779.764966] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.778574] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527507f5-b8b4-9c45-92db-e5664b0e537d, 'name': SearchDatastore_Task, 'duration_secs': 0.010145} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.778890] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.779159] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.779393] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.779537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.779713] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.779970] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-328f76b2-e6dd-4d77-8d59-ce833cc5d953 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.787487] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.787681] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.788476] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-716a92d2-7f98-45c1-ba6f-c235a9ff50e8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.796125] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 779.796125] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c2d752-4606-10ef-5d20-e8b42b7a53fe" [ 779.796125] env[61936]: _type = "Task" [ 779.796125] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.803206] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c2d752-4606-10ef-5d20-e8b42b7a53fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.903502] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.140024] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.140024] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.306964] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c2d752-4606-10ef-5d20-e8b42b7a53fe, 'name': SearchDatastore_Task, 'duration_secs': 0.00867} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.311240] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7152f078-e089-44c4-ae1d-83a846066ca2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.317550] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 780.317550] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523549b8-bf02-c111-6e8a-f0ef791e6f88" [ 780.317550] env[61936]: _type = "Task" [ 780.317550] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.327719] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523549b8-bf02-c111-6e8a-f0ef791e6f88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.398644] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "aa37292e-8499-46aa-8021-cb8d88b5a35a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.398883] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "aa37292e-8499-46aa-8021-cb8d88b5a35a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.406096] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Releasing lock "refresh_cache-03ab71e9-64b7-4534-a693-473f3903c511" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.406534] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 780.406733] env[61936]: DEBUG nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 780.406891] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.441774] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.566520] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95b7ff3-f15c-4f4d-b4ab-058d884a3bb5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.574496] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568870f8-d3e8-457d-9327-e14740a17de5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.607310] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f82442-c525-4364-b719-f61ed7a6b63a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.614893] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824101a8-3457-4af5-8000-92f5277319d3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.630245] env[61936]: DEBUG nova.compute.provider_tree [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.828433] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523549b8-bf02-c111-6e8a-f0ef791e6f88, 'name': SearchDatastore_Task, 'duration_secs': 0.038212} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.828736] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.828998] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.829273] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-85cf6dc1-0cfc-4f39-b04d-563b02fcbf32 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.837494] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 780.837494] env[61936]: value = "task-1252773" [ 780.837494] env[61936]: _type = "Task" [ 780.837494] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.849697] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252773, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.948654] env[61936]: DEBUG nova.network.neutron [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.133588] env[61936]: DEBUG nova.scheduler.client.report [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 781.354425] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252773, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.453980] env[61936]: INFO nova.compute.manager [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] [instance: 03ab71e9-64b7-4534-a693-473f3903c511] Took 1.04 seconds to deallocate network for instance. [ 781.639693] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.640287] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 781.644804] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.840s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.646452] env[61936]: INFO nova.compute.claims [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.851806] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252773, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.683847} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.852147] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 781.852329] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.852711] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84f71e68-a94c-4f76-901d-ba5a95881ff6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.860643] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 781.860643] env[61936]: value = "task-1252774" [ 781.860643] env[61936]: _type = "Task" [ 781.860643] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.870552] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252774, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.146759] env[61936]: DEBUG nova.compute.utils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.148459] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 782.148459] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 782.219317] env[61936]: DEBUG nova.policy [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5fe2328df734cea9dca08978fba8134', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edc02335db4d4f60a9ed1235dacdb348', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 782.371458] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252774, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067226} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.371795] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 782.372817] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af841cb2-84ae-4d33-956f-0a55cdac427f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.394749] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 782.395041] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-497a3715-6ebb-4a1c-aa2d-795e7e92f03e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.414402] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 782.414402] env[61936]: value = "task-1252775" [ 782.414402] env[61936]: _type = "Task" [ 782.414402] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.424075] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252775, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.512716] env[61936]: INFO nova.scheduler.client.report [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Deleted allocations for instance 03ab71e9-64b7-4534-a693-473f3903c511 [ 782.523651] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Successfully created port: 781d1814-e806-4aeb-87d9-ed6a427f2194 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.651061] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 782.773597] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.773802] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.925957] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252775, 'name': ReconfigVM_Task, 'duration_secs': 0.273742} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.926484] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.926896] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a5b1226-17df-4554-8b02-a2b8e6ec9d13 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.933325] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 782.933325] env[61936]: value = "task-1252776" [ 782.933325] env[61936]: _type = "Task" [ 782.933325] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.945073] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252776, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.028739] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf1ebb71-0110-4fa7-84fb-a7165a32a7cd tempest-ServerActionsTestJSON-317992167 tempest-ServerActionsTestJSON-317992167-project-member] Lock "03ab71e9-64b7-4534-a693-473f3903c511" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.034s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.034022] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c935021b-a05f-4223-8a5d-362f1dc555a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.039543] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc337571-3150-4210-95aa-aa06efbd178b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.073374] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed39d52-a59f-4aa5-ad55-0beba5f04b53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.081132] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4ab2e7-cd2e-4d9a-884f-b01832c4407e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.097733] env[61936]: DEBUG nova.compute.provider_tree [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.290634] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.290814] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 783.290930] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Rebuilding the list of instances to heal {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10330}} [ 783.448912] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252776, 'name': Rename_Task, 'duration_secs': 0.129876} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.448912] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 783.448912] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61082fd4-2c47-4975-a262-9981065ff0c5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.456177] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 783.456177] env[61936]: value = "task-1252777" [ 783.456177] env[61936]: _type = "Task" [ 783.456177] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.462162] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.510106] env[61936]: DEBUG nova.compute.manager [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Received event network-changed-781d1814-e806-4aeb-87d9-ed6a427f2194 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 783.510106] env[61936]: DEBUG nova.compute.manager [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Refreshing instance network info cache due to event network-changed-781d1814-e806-4aeb-87d9-ed6a427f2194. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 783.510106] env[61936]: DEBUG oslo_concurrency.lockutils [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] Acquiring lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.510106] env[61936]: DEBUG oslo_concurrency.lockutils [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] Acquired lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.510106] env[61936]: DEBUG nova.network.neutron [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Refreshing network info cache for port 781d1814-e806-4aeb-87d9-ed6a427f2194 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 783.537643] env[61936]: DEBUG nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 783.623303] env[61936]: ERROR nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [req-42b51206-7122-480b-a906-c2628d2a0e96] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID abd382a7-b427-4dcc-a050-3fb56fdd60ac. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-42b51206-7122-480b-a906-c2628d2a0e96"}]} [ 783.640478] env[61936]: DEBUG nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Refreshing inventories for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 783.659589] env[61936]: DEBUG nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Updating ProviderTree inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 783.659589] env[61936]: DEBUG nova.compute.provider_tree [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.662777] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 783.675401] env[61936]: DEBUG nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Refreshing aggregate associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, aggregates: None {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 783.696336] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 783.696336] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.696336] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 783.696520] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.696520] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 783.696520] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 783.696520] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 783.696520] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 783.696775] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 783.696775] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 783.696775] env[61936]: DEBUG nova.virt.hardware [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 783.697117] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90d7aab-51a4-4111-9938-b1203c07bee3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.701280] env[61936]: DEBUG nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Refreshing trait associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 783.712526] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43030da1-5049-4ea4-99ae-da4e41e87484 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.754940] env[61936]: ERROR nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. [ 783.754940] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 783.754940] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 783.754940] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 783.754940] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.754940] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 783.754940] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.754940] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 783.754940] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.754940] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 783.754940] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.754940] env[61936]: ERROR nova.compute.manager raise self.value [ 783.754940] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.754940] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 783.754940] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.754940] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 783.757696] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.757696] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 783.757696] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. [ 783.757696] env[61936]: ERROR nova.compute.manager [ 783.757696] env[61936]: Traceback (most recent call last): [ 783.757696] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 783.757696] env[61936]: listener.cb(fileno) [ 783.757696] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.757696] env[61936]: result = function(*args, **kwargs) [ 783.757696] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 783.757696] env[61936]: return func(*args, **kwargs) [ 783.757696] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 783.757696] env[61936]: raise e [ 783.757696] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 783.757696] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 783.757696] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.757696] env[61936]: created_port_ids = self._update_ports_for_instance( [ 783.757696] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.757696] env[61936]: with excutils.save_and_reraise_exception(): [ 783.757696] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.757696] env[61936]: self.force_reraise() [ 783.757696] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.757696] env[61936]: raise self.value [ 783.757696] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.757696] env[61936]: updated_port = self._update_port( [ 783.757696] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.757696] env[61936]: _ensure_no_port_binding_failure(port) [ 783.757696] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.757696] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 783.758733] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. [ 783.758733] env[61936]: Removing descriptor: 15 [ 783.758733] env[61936]: ERROR nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Traceback (most recent call last): [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] yield resources [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self.driver.spawn(context, instance, image_meta, [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.758733] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] vm_ref = self.build_virtual_machine(instance, [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] for vif in network_info: [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return self._sync_wrapper(fn, *args, **kwargs) [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self.wait() [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self[:] = self._gt.wait() [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return self._exit_event.wait() [ 783.759255] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] result = hub.switch() [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return self.greenlet.switch() [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] result = function(*args, **kwargs) [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return func(*args, **kwargs) [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] raise e [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] nwinfo = self.network_api.allocate_for_instance( [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.759675] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] created_port_ids = self._update_ports_for_instance( [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] with excutils.save_and_reraise_exception(): [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self.force_reraise() [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] raise self.value [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] updated_port = self._update_port( [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] _ensure_no_port_binding_failure(port) [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.760078] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] raise exception.PortBindingFailed(port_id=port['id']) [ 783.760441] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] nova.exception.PortBindingFailed: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. [ 783.760441] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] [ 783.760441] env[61936]: INFO nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Terminating instance [ 783.794715] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 783.794715] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 783.794715] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 783.794904] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 783.794945] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 783.794945] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Didn't find any instances for network info cache update. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10412}} [ 783.795240] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.795395] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.795535] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.795675] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.795809] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.796022] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.797793] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 783.797793] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 783.969847] env[61936]: DEBUG oslo_vmware.api [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252777, 'name': PowerOnVM_Task, 'duration_secs': 0.46547} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.970139] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.970484] env[61936]: INFO nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Took 4.83 seconds to spawn the instance on the hypervisor. [ 783.970650] env[61936]: DEBUG nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 783.971454] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1d3447-b5f5-4670-bd50-3b5cb1b822da {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.055774] env[61936]: DEBUG nova.network.neutron [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.060674] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.099110] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e775332e-d78a-45c6-adc6-ef8bb6cc863a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.108668] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8baf4996-5d7c-4312-9be7-f9f840f9d653 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.141913] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a0685b-a6cd-4e86-aae7-ea6cf66d4bc9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.151442] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304317cd-1d72-4262-a887-a2423c88e79f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.165548] env[61936]: DEBUG nova.compute.provider_tree [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.192079] env[61936]: DEBUG nova.network.neutron [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.263966] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.299088] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.493738] env[61936]: INFO nova.compute.manager [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Took 24.00 seconds to build instance. [ 784.695463] env[61936]: DEBUG oslo_concurrency.lockutils [req-47ac2434-2c5e-468b-beb7-10fe3cb17f72 req-af1d1e58-f232-445a-bad8-e5047a76f611 service nova] Releasing lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.696026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquired lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.696090] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.710733] env[61936]: DEBUG nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 784.710926] env[61936]: DEBUG nova.compute.provider_tree [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 78 to 79 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 784.711121] env[61936]: DEBUG nova.compute.provider_tree [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.998347] env[61936]: DEBUG oslo_concurrency.lockutils [None req-df45cb5c-5a86-4797-8f83-ebc61c59de9c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.338s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.216230] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.571s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.216973] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 785.221039] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.817s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.230370] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.416624] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.500771] env[61936]: DEBUG nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 785.681544] env[61936]: INFO nova.compute.manager [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Rebuilding instance [ 785.730030] env[61936]: DEBUG nova.compute.utils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.731732] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 785.731901] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 785.763657] env[61936]: DEBUG nova.compute.manager [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 785.765647] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a625b6-e5ae-4875-9a24-2a946743b7cc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.811176] env[61936]: DEBUG nova.policy [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5424d52a4bd5486692a74f1f7f5447c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '618a7dd2288842efbbe14291df828550', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.920370] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Releasing lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.921789] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 785.921789] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.921789] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec9bd1f7-8c5a-4026-b0b4-29088ef7d781 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.932163] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.932465] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.941594] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b368d0-5767-46e6-829e-0a971a023ae3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.961182] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.961416] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.970175] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85 could not be found. [ 785.970175] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.970175] env[61936]: INFO nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Took 0.05 seconds to destroy the instance on the hypervisor. [ 785.970175] env[61936]: DEBUG oslo.service.loopingcall [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.970175] env[61936]: DEBUG nova.compute.manager [-] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 785.970175] env[61936]: DEBUG nova.network.neutron [-] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 786.023970] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.035798] env[61936]: DEBUG nova.network.neutron [-] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.062541] env[61936]: DEBUG nova.compute.manager [req-f8caaa13-dd65-451b-a3d9-3fee81af65dd req-e8711d7b-8792-4384-801c-c995939597fb service nova] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Received event network-vif-deleted-781d1814-e806-4aeb-87d9-ed6a427f2194 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 786.220038] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffa0bc3-a110-4a5f-a587-b0ad96a2d14e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.228119] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c45180-2efe-4af1-ae9e-ff670b5d3f52 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.258380] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 786.263324] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d464bf-88ad-485c-98d4-2d270a211e7e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.273526] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d2bb61-b2cc-4299-a022-1767d96e8376 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.287897] env[61936]: DEBUG nova.compute.provider_tree [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.333886] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Successfully created port: df6997b1-d103-47ef-bb00-0393194e7d85 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.546824] env[61936]: DEBUG nova.network.neutron [-] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.791265] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.792827] env[61936]: DEBUG nova.scheduler.client.report [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 786.797269] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5ee059c-f12f-4eb7-8f39-edb64590832c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.808343] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 786.808343] env[61936]: value = "task-1252778" [ 786.808343] env[61936]: _type = "Task" [ 786.808343] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.820378] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.050031] env[61936]: INFO nova.compute.manager [-] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Took 1.08 seconds to deallocate network for instance. [ 787.052402] env[61936]: DEBUG nova.compute.claims [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 787.052579] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.274497] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 787.301452] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 787.301673] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.301820] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 787.301988] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.302148] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 787.302286] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 787.302481] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 787.302628] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 787.302782] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 787.302932] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 787.303106] env[61936]: DEBUG nova.virt.hardware [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 787.303820] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.083s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.304459] env[61936]: ERROR nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Traceback (most recent call last): [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self.driver.spawn(context, instance, image_meta, [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] vm_ref = self.build_virtual_machine(instance, [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.304459] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] for vif in network_info: [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return self._sync_wrapper(fn, *args, **kwargs) [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self.wait() [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self[:] = self._gt.wait() [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return self._exit_event.wait() [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] result = hub.switch() [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 787.304865] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return self.greenlet.switch() [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] result = function(*args, **kwargs) [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] return func(*args, **kwargs) [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] raise e [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] nwinfo = self.network_api.allocate_for_instance( [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] created_port_ids = self._update_ports_for_instance( [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] with excutils.save_and_reraise_exception(): [ 787.306395] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] self.force_reraise() [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] raise self.value [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] updated_port = self._update_port( [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] _ensure_no_port_binding_failure(port) [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] raise exception.PortBindingFailed(port_id=port['id']) [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] nova.exception.PortBindingFailed: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. [ 787.306828] env[61936]: ERROR nova.compute.manager [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] [ 787.307199] env[61936]: DEBUG nova.compute.utils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 787.307199] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483e6e6d-0efd-4d26-a79c-ca160dbb2cfa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.309608] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Build of instance 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca was re-scheduled: Binding failed for port 5c4ef559-9891-4f7f-8608-19a8fbd74016, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 787.310158] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 787.310381] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Acquiring lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.310515] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Acquired lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.310662] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.311637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.526s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.313020] env[61936]: INFO nova.compute.claims [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.327062] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252778, 'name': PowerOffVM_Task, 'duration_secs': 0.217689} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.327661] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 787.327949] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 787.329165] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7bd223-35f7-4f2b-b4f5-c97113bd0a77 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.333534] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179e329f-d948-4a22-b48c-101216fd437b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.348517] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 787.348754] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c71b3010-c246-48e9-ba2b-50b99a145d24 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.371144] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 787.371389] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 787.371565] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Deleting the datastore file [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.372104] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28430606-7a9c-42ed-9f91-4f7aec349d20 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.381225] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 787.381225] env[61936]: value = "task-1252780" [ 787.381225] env[61936]: _type = "Task" [ 787.381225] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.389869] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.552739] env[61936]: ERROR nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. [ 787.552739] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 787.552739] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 787.552739] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 787.552739] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.552739] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 787.552739] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.552739] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 787.552739] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.552739] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 787.552739] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.552739] env[61936]: ERROR nova.compute.manager raise self.value [ 787.552739] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.552739] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 787.552739] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.552739] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 787.553635] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.553635] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 787.553635] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. [ 787.553635] env[61936]: ERROR nova.compute.manager [ 787.553635] env[61936]: Traceback (most recent call last): [ 787.553635] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 787.553635] env[61936]: listener.cb(fileno) [ 787.553635] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.553635] env[61936]: result = function(*args, **kwargs) [ 787.553635] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 787.553635] env[61936]: return func(*args, **kwargs) [ 787.553635] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 787.553635] env[61936]: raise e [ 787.553635] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 787.553635] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 787.553635] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.553635] env[61936]: created_port_ids = self._update_ports_for_instance( [ 787.553635] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.553635] env[61936]: with excutils.save_and_reraise_exception(): [ 787.553635] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.553635] env[61936]: self.force_reraise() [ 787.553635] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.553635] env[61936]: raise self.value [ 787.553635] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.553635] env[61936]: updated_port = self._update_port( [ 787.553635] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.553635] env[61936]: _ensure_no_port_binding_failure(port) [ 787.553635] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.553635] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 787.554994] env[61936]: nova.exception.PortBindingFailed: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. [ 787.554994] env[61936]: Removing descriptor: 15 [ 787.554994] env[61936]: ERROR nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] Traceback (most recent call last): [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] yield resources [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self.driver.spawn(context, instance, image_meta, [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.554994] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] vm_ref = self.build_virtual_machine(instance, [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] for vif in network_info: [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return self._sync_wrapper(fn, *args, **kwargs) [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self.wait() [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self[:] = self._gt.wait() [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return self._exit_event.wait() [ 787.555632] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] result = hub.switch() [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return self.greenlet.switch() [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] result = function(*args, **kwargs) [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return func(*args, **kwargs) [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] raise e [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] nwinfo = self.network_api.allocate_for_instance( [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.556168] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] created_port_ids = self._update_ports_for_instance( [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] with excutils.save_and_reraise_exception(): [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self.force_reraise() [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] raise self.value [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] updated_port = self._update_port( [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] _ensure_no_port_binding_failure(port) [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.556736] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] raise exception.PortBindingFailed(port_id=port['id']) [ 787.557226] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] nova.exception.PortBindingFailed: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. [ 787.557226] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] [ 787.557226] env[61936]: INFO nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Terminating instance [ 787.846856] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.890867] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.084301} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.893393] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.893580] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.893755] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.940282] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.061321] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Acquiring lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.061520] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Acquired lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.061714] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.093251] env[61936]: DEBUG nova.compute.manager [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Received event network-changed-df6997b1-d103-47ef-bb00-0393194e7d85 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 788.093251] env[61936]: DEBUG nova.compute.manager [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Refreshing instance network info cache due to event network-changed-df6997b1-d103-47ef-bb00-0393194e7d85. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 788.093433] env[61936]: DEBUG oslo_concurrency.lockutils [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] Acquiring lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.443258] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Releasing lock "refresh_cache-8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.443258] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 788.443258] env[61936]: DEBUG nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 788.443480] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.461589] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.579295] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.646212] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0c7b20-1e2a-48f7-8e22-aef8b21dadb8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.654085] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54ffeea-0fec-4623-90f1-7768f4222720 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.683976] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de83aa34-ed18-4c13-aa64-8f4678e6a29b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.691208] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cb1621-ab54-461d-b3f6-50acb49f0ef4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.703952] env[61936]: DEBUG nova.compute.provider_tree [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.705631] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.929482] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 788.929883] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.930176] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 788.930488] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.930830] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 788.931100] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 788.931465] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 788.931749] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 788.932068] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 788.932322] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 788.932509] env[61936]: DEBUG nova.virt.hardware [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 788.933388] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6a3fd7-cc87-46bc-89a9-d3d0f2adc8b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.941185] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1ac5c6-3257-4d04-b55b-0e4ab4e9936a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.955132] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.960578] env[61936]: DEBUG oslo.service.loopingcall [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.960788] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.960983] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f7e3438-9922-43d5-84fe-fe9eca54b5cc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.972174] env[61936]: DEBUG nova.network.neutron [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.978821] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.978821] env[61936]: value = "task-1252781" [ 788.978821] env[61936]: _type = "Task" [ 788.978821] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.985870] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252781, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.208691] env[61936]: DEBUG nova.scheduler.client.report [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 789.213307] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Releasing lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.213503] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 789.213754] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.214337] env[61936]: DEBUG oslo_concurrency.lockutils [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] Acquired lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.214566] env[61936]: DEBUG nova.network.neutron [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Refreshing network info cache for port df6997b1-d103-47ef-bb00-0393194e7d85 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.215612] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-684a2af9-e088-437e-beb5-fdc0f8923e99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.226805] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad10b33-1221-4315-a8df-021147f3c23c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.250825] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4755f16d-254b-47e9-8b50-05df95690445 could not be found. [ 789.250929] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.251164] env[61936]: INFO nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Took 0.04 seconds to destroy the instance on the hypervisor. [ 789.251447] env[61936]: DEBUG oslo.service.loopingcall [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.251704] env[61936]: DEBUG nova.compute.manager [-] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 789.251808] env[61936]: DEBUG nova.network.neutron [-] [instance: 4755f16d-254b-47e9-8b50-05df95690445] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.265670] env[61936]: DEBUG nova.network.neutron [-] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.475490] env[61936]: INFO nova.compute.manager [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] [instance: 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca] Took 1.03 seconds to deallocate network for instance. [ 789.488234] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252781, 'name': CreateVM_Task, 'duration_secs': 0.260551} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.488393] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.488798] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.489022] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.489332] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.489560] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c52f2ff2-3c4c-48f0-9a4b-e6c5928fb61c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.493792] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 789.493792] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526a43c5-45ee-16cc-4009-d540736db892" [ 789.493792] env[61936]: _type = "Task" [ 789.493792] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.501160] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526a43c5-45ee-16cc-4009-d540736db892, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.715696] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.716292] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 789.719219] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.095s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.720774] env[61936]: INFO nova.compute.claims [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.743606] env[61936]: DEBUG nova.network.neutron [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.767809] env[61936]: DEBUG nova.network.neutron [-] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.850122] env[61936]: DEBUG nova.network.neutron [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.004194] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526a43c5-45ee-16cc-4009-d540736db892, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.005440] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.005440] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.005440] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.005440] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.005694] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.005694] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-297a51f6-5c32-40a1-baf8-605ec3d5a95c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.014048] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.014236] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.014934] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b80f3a7-4a43-4fba-a5e3-3f0d91bfd4b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.022268] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 790.022268] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523b302b-f820-4806-c538-0dee7523d6f6" [ 790.022268] env[61936]: _type = "Task" [ 790.022268] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.029549] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523b302b-f820-4806-c538-0dee7523d6f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.229655] env[61936]: DEBUG nova.compute.utils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.231237] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 790.231417] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.273824] env[61936]: INFO nova.compute.manager [-] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Took 1.02 seconds to deallocate network for instance. [ 790.276398] env[61936]: DEBUG nova.compute.claims [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 790.276579] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.288177] env[61936]: DEBUG nova.policy [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5eb3b92a98994ab5a6371f5377d56c88', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b7ed656a83743ddb23b452c38d5b607', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.354639] env[61936]: DEBUG oslo_concurrency.lockutils [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] Releasing lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.354639] env[61936]: DEBUG nova.compute.manager [req-fc963d7c-1606-4840-b4b7-d9d9c3708bbb req-ab35b901-f85e-4257-8712-745c5f149c74 service nova] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Received event network-vif-deleted-df6997b1-d103-47ef-bb00-0393194e7d85 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 790.511492] env[61936]: INFO nova.scheduler.client.report [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Deleted allocations for instance 8bfe1de5-1a37-49f4-997e-19ae5c33d8ca [ 790.534135] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523b302b-f820-4806-c538-0dee7523d6f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009061} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.535502] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bb5312e-0947-408c-9a6b-a78116ddc8b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.541185] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 790.541185] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521aacf6-779a-f110-7955-3d398dc95442" [ 790.541185] env[61936]: _type = "Task" [ 790.541185] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.549153] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521aacf6-779a-f110-7955-3d398dc95442, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.734691] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 790.779035] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Successfully created port: 462018e4-3904-4c95-8328-7819e8709198 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.022590] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b904f828-bacf-433d-b206-18ff26b31ea6 tempest-ServersNegativeTestMultiTenantJSON-1579216929 tempest-ServersNegativeTestMultiTenantJSON-1579216929-project-member] Lock "8bfe1de5-1a37-49f4-997e-19ae5c33d8ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.369s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.055729] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521aacf6-779a-f110-7955-3d398dc95442, 'name': SearchDatastore_Task, 'duration_secs': 0.009602} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.056695] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.056695] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.056695] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0932cec8-9696-405e-b300-8ea0cd7715bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.066457] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 791.066457] env[61936]: value = "task-1252782" [ 791.066457] env[61936]: _type = "Task" [ 791.066457] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.078863] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252782, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.098780] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35434c8a-64e6-43a3-8d48-92d8893b9047 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.106362] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfce99d-0c4f-4c97-8ab5-89fc16e306a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.139369] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4df2a28-2b7c-46cd-8438-1a3fc6e3605e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.146804] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7d4855-a869-499c-b366-93ab68aeacba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.160039] env[61936]: DEBUG nova.compute.provider_tree [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.527878] env[61936]: DEBUG nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 791.580763] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252782, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444355} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.581026] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.581695] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.581695] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f8da390-d083-415f-be88-ae9b647ac441 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.588042] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 791.588042] env[61936]: value = "task-1252783" [ 791.588042] env[61936]: _type = "Task" [ 791.588042] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.595651] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252783, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.662760] env[61936]: DEBUG nova.compute.manager [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Received event network-changed-462018e4-3904-4c95-8328-7819e8709198 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 791.662952] env[61936]: DEBUG nova.compute.manager [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Refreshing instance network info cache due to event network-changed-462018e4-3904-4c95-8328-7819e8709198. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 791.663190] env[61936]: DEBUG oslo_concurrency.lockutils [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] Acquiring lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.663326] env[61936]: DEBUG oslo_concurrency.lockutils [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] Acquired lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.663477] env[61936]: DEBUG nova.network.neutron [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Refreshing network info cache for port 462018e4-3904-4c95-8328-7819e8709198 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.665528] env[61936]: DEBUG nova.scheduler.client.report [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 791.751198] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 791.778180] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 791.778432] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.778578] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 791.778745] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.778880] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 791.779027] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 791.780678] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 791.780678] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 791.780678] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 791.780678] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 791.780678] env[61936]: DEBUG nova.virt.hardware [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 791.780891] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4a524d-311a-4660-9ebb-8a658032a958 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.788953] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf825f4-05cf-4207-b1dd-df7532edbcef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.053417] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.064047] env[61936]: ERROR nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. [ 792.064047] env[61936]: ERROR nova.compute.manager Traceback (most recent call last): [ 792.064047] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 792.064047] env[61936]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 792.064047] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.064047] env[61936]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 792.064047] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.064047] env[61936]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 792.064047] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.064047] env[61936]: ERROR nova.compute.manager self.force_reraise() [ 792.064047] env[61936]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.064047] env[61936]: ERROR nova.compute.manager raise self.value [ 792.064047] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.064047] env[61936]: ERROR nova.compute.manager updated_port = self._update_port( [ 792.064047] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.064047] env[61936]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 792.064530] env[61936]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.064530] env[61936]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 792.064530] env[61936]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. [ 792.064530] env[61936]: ERROR nova.compute.manager [ 792.064530] env[61936]: Traceback (most recent call last): [ 792.064530] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 792.064530] env[61936]: listener.cb(fileno) [ 792.064530] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.064530] env[61936]: result = function(*args, **kwargs) [ 792.064530] env[61936]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.064530] env[61936]: return func(*args, **kwargs) [ 792.064530] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 792.064530] env[61936]: raise e [ 792.064530] env[61936]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 792.064530] env[61936]: nwinfo = self.network_api.allocate_for_instance( [ 792.064530] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.064530] env[61936]: created_port_ids = self._update_ports_for_instance( [ 792.064530] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.064530] env[61936]: with excutils.save_and_reraise_exception(): [ 792.064530] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.064530] env[61936]: self.force_reraise() [ 792.064530] env[61936]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.064530] env[61936]: raise self.value [ 792.064530] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.064530] env[61936]: updated_port = self._update_port( [ 792.064530] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.064530] env[61936]: _ensure_no_port_binding_failure(port) [ 792.064530] env[61936]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.064530] env[61936]: raise exception.PortBindingFailed(port_id=port['id']) [ 792.065350] env[61936]: nova.exception.PortBindingFailed: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. [ 792.065350] env[61936]: Removing descriptor: 15 [ 792.065350] env[61936]: ERROR nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Traceback (most recent call last): [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] yield resources [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self.driver.spawn(context, instance, image_meta, [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.065350] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] vm_ref = self.build_virtual_machine(instance, [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] for vif in network_info: [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return self._sync_wrapper(fn, *args, **kwargs) [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self.wait() [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self[:] = self._gt.wait() [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return self._exit_event.wait() [ 792.065710] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] result = hub.switch() [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return self.greenlet.switch() [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] result = function(*args, **kwargs) [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return func(*args, **kwargs) [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] raise e [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] nwinfo = self.network_api.allocate_for_instance( [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.066163] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] created_port_ids = self._update_ports_for_instance( [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] with excutils.save_and_reraise_exception(): [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self.force_reraise() [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] raise self.value [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] updated_port = self._update_port( [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] _ensure_no_port_binding_failure(port) [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.066534] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] raise exception.PortBindingFailed(port_id=port['id']) [ 792.066892] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] nova.exception.PortBindingFailed: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. [ 792.066892] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] [ 792.066892] env[61936]: INFO nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Terminating instance [ 792.101101] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252783, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069108} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.101101] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.101101] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4fa795-a332-4276-9b98-d1b986b27d93 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.121376] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.121917] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74719399-c78c-4bdf-837a-eb424de13746 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.141123] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 792.141123] env[61936]: value = "task-1252784" [ 792.141123] env[61936]: _type = "Task" [ 792.141123] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.149036] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252784, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.173945] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.174465] env[61936]: DEBUG nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 792.176959] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.171s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.189666] env[61936]: DEBUG nova.network.neutron [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.260709] env[61936]: DEBUG nova.network.neutron [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.572440] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.652797] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252784, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.682373] env[61936]: DEBUG nova.compute.utils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 792.687545] env[61936]: DEBUG nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 792.687729] env[61936]: DEBUG nova.network.neutron [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 792.756713] env[61936]: DEBUG nova.policy [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5eb3b92a98994ab5a6371f5377d56c88', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b7ed656a83743ddb23b452c38d5b607', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.763325] env[61936]: DEBUG oslo_concurrency.lockutils [req-164bc4b4-9afd-439a-8245-edc0aa45e32d req-204d25a6-af85-4d18-8e45-222de507073a service nova] Releasing lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.765764] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquired lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.765951] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.033724] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c46612-db69-4490-a64b-b30ae516f21a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.041675] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bd1f8e-174c-4306-a2be-ba272d5e7d30 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.074980] env[61936]: DEBUG nova.network.neutron [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Successfully created port: 1056fca0-7eeb-464b-befc-9a19819b7a58 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.078741] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96244d3-8352-4605-ac51-5ede4a91d67a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.085062] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878e2b47-9e83-4620-847f-f1a169cd1c6a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.098540] env[61936]: DEBUG nova.compute.provider_tree [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.151408] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252784, 'name': ReconfigVM_Task, 'duration_secs': 0.525238} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.151677] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a/498a77d6-ca0c-4841-b4a2-b7a024281c6a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.152306] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aaa482b7-b075-41c0-9dc7-8814431cc58b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.157867] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 793.157867] env[61936]: value = "task-1252785" [ 793.157867] env[61936]: _type = "Task" [ 793.157867] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.165328] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252785, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.188574] env[61936]: DEBUG nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 793.295275] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.414841] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.604417] env[61936]: DEBUG nova.scheduler.client.report [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 793.668881] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252785, 'name': Rename_Task, 'duration_secs': 0.144499} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.669368] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 793.669717] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46d27180-908f-41e8-9c0e-13801a4a3bab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.677022] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 793.677022] env[61936]: value = "task-1252786" [ 793.677022] env[61936]: _type = "Task" [ 793.677022] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.684813] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252786, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.723508] env[61936]: DEBUG nova.compute.manager [req-02735765-3bd3-4425-be9d-492c382f430c req-6f41be85-d9b4-4fea-bebd-538ac080b14d service nova] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Received event network-vif-deleted-462018e4-3904-4c95-8328-7819e8709198 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 793.919119] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Releasing lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.919573] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 793.919755] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.920075] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-49ea1299-c97e-447a-a3de-84ab7e9d9882 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.929819] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe746e03-93c1-42ce-899e-1d8ca4c14770 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.953031] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2fdb605e-72d5-4eef-bab2-0601e730243a could not be found. [ 793.953274] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 793.953457] env[61936]: INFO nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 793.953703] env[61936]: DEBUG oslo.service.loopingcall [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.953916] env[61936]: DEBUG nova.compute.manager [-] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 793.954013] env[61936]: DEBUG nova.network.neutron [-] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.969721] env[61936]: DEBUG nova.network.neutron [-] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.110918] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.111571] env[61936]: ERROR nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Traceback (most recent call last): [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self.driver.spawn(context, instance, image_meta, [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] vm_ref = self.build_virtual_machine(instance, [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] vif_infos = vmwarevif.get_vif_info(self._session, [ 794.111571] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] for vif in network_info: [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] return self._sync_wrapper(fn, *args, **kwargs) [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self.wait() [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self[:] = self._gt.wait() [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] return self._exit_event.wait() [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] current.throw(*self._exc) [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 794.112171] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] result = function(*args, **kwargs) [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] return func(*args, **kwargs) [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] raise e [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] nwinfo = self.network_api.allocate_for_instance( [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] created_port_ids = self._update_ports_for_instance( [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] with excutils.save_and_reraise_exception(): [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] self.force_reraise() [ 794.112538] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] raise self.value [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] updated_port = self._update_port( [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] _ensure_no_port_binding_failure(port) [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] raise exception.PortBindingFailed(port_id=port['id']) [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] nova.exception.PortBindingFailed: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. [ 794.112924] env[61936]: ERROR nova.compute.manager [instance: 0b14583c-dd0f-429d-a386-2acda353465a] [ 794.112924] env[61936]: DEBUG nova.compute.utils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 794.113914] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.578s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.118579] env[61936]: INFO nova.compute.claims [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.121178] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Build of instance 0b14583c-dd0f-429d-a386-2acda353465a was re-scheduled: Binding failed for port 6fc5c96c-3d2d-46da-9375-74c69c0d59d7, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 794.121633] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 794.121851] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Acquiring lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.121996] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Acquired lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.122185] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.186677] env[61936]: DEBUG oslo_vmware.api [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252786, 'name': PowerOnVM_Task, 'duration_secs': 0.432951} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.186961] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 794.187192] env[61936]: DEBUG nova.compute.manager [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 794.188125] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d84383a-354e-4943-a7eb-577d8e75a5a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.203028] env[61936]: DEBUG nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 794.227757] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 794.228039] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.228200] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 794.228375] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.228515] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 794.228653] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 794.228848] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 794.229053] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 794.229267] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 794.229384] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 794.229546] env[61936]: DEBUG nova.virt.hardware [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 794.230750] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40344f7-3a17-4521-ae1c-218a566f8fa9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.239097] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224ba0ba-ef12-4fb4-a712-57b060130ebd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.475825] env[61936]: DEBUG nova.network.neutron [-] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.662479] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.706190] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.877539] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.908874] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.909232] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.909482] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.909698] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.909902] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.911915] env[61936]: INFO nova.compute.manager [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Terminating instance [ 794.978889] env[61936]: INFO nova.compute.manager [-] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Took 1.02 seconds to deallocate network for instance. [ 794.981469] env[61936]: DEBUG nova.compute.claims [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Aborting claim: {{(pid=61936) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 794.981730] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.101801] env[61936]: DEBUG nova.network.neutron [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Successfully updated port: 1056fca0-7eeb-464b-befc-9a19819b7a58 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.379871] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Releasing lock "refresh_cache-0b14583c-dd0f-429d-a386-2acda353465a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.380117] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 795.380624] env[61936]: DEBUG nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 795.380624] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.407521] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.416859] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "refresh_cache-498a77d6-ca0c-4841-b4a2-b7a024281c6a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.417110] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquired lock "refresh_cache-498a77d6-ca0c-4841-b4a2-b7a024281c6a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.417310] env[61936]: DEBUG nova.network.neutron [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.451970] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff54577-dde9-447b-9f7b-0fefd65be107 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.459907] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c546535a-fa79-4a8c-9045-af81bca455b6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.490605] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234b34dd-b0c8-418f-97df-7cb72cf24537 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.497966] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0da01c-761b-4905-90e5-f567ab80c6ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.512046] env[61936]: DEBUG nova.compute.provider_tree [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.604595] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "refresh_cache-50cb3f79-39b2-46ab-a35c-fdad3bb266cd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.604744] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquired lock "refresh_cache-50cb3f79-39b2-46ab-a35c-fdad3bb266cd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.604894] env[61936]: DEBUG nova.network.neutron [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.749910] env[61936]: DEBUG nova.compute.manager [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Received event network-vif-plugged-1056fca0-7eeb-464b-befc-9a19819b7a58 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 795.749910] env[61936]: DEBUG oslo_concurrency.lockutils [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] Acquiring lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.750320] env[61936]: DEBUG oslo_concurrency.lockutils [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.750320] env[61936]: DEBUG oslo_concurrency.lockutils [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.750529] env[61936]: DEBUG nova.compute.manager [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] No waiting events found dispatching network-vif-plugged-1056fca0-7eeb-464b-befc-9a19819b7a58 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 795.751387] env[61936]: WARNING nova.compute.manager [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Received unexpected event network-vif-plugged-1056fca0-7eeb-464b-befc-9a19819b7a58 for instance with vm_state building and task_state spawning. [ 795.751387] env[61936]: DEBUG nova.compute.manager [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Received event network-changed-1056fca0-7eeb-464b-befc-9a19819b7a58 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 795.752025] env[61936]: DEBUG nova.compute.manager [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Refreshing instance network info cache due to event network-changed-1056fca0-7eeb-464b-befc-9a19819b7a58. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 795.752025] env[61936]: DEBUG oslo_concurrency.lockutils [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] Acquiring lock "refresh_cache-50cb3f79-39b2-46ab-a35c-fdad3bb266cd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.910059] env[61936]: DEBUG nova.network.neutron [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.940233] env[61936]: DEBUG nova.network.neutron [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.989916] env[61936]: DEBUG nova.network.neutron [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.014233] env[61936]: DEBUG nova.scheduler.client.report [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 796.147062] env[61936]: DEBUG nova.network.neutron [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.393378] env[61936]: DEBUG nova.network.neutron [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Updating instance_info_cache with network_info: [{"id": "1056fca0-7eeb-464b-befc-9a19819b7a58", "address": "fa:16:3e:cf:5b:c7", "network": {"id": "341c94d2-35e4-44d5-b231-2e36201f4ef7", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "0a2bb4cb48fe49aba25dc72f9f3195d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1056fca0-7e", "ovs_interfaceid": "1056fca0-7eeb-464b-befc-9a19819b7a58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.412859] env[61936]: INFO nova.compute.manager [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] [instance: 0b14583c-dd0f-429d-a386-2acda353465a] Took 1.03 seconds to deallocate network for instance. [ 796.492708] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Releasing lock "refresh_cache-498a77d6-ca0c-4841-b4a2-b7a024281c6a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.493432] env[61936]: DEBUG nova.compute.manager [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 796.493676] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.494564] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9509fa0-5f80-43de-bd70-808da29b2d00 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.502209] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.502438] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9b114ac-c7b8-4b01-9724-88d2173708bb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.509232] env[61936]: DEBUG oslo_vmware.api [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 796.509232] env[61936]: value = "task-1252787" [ 796.509232] env[61936]: _type = "Task" [ 796.509232] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.516866] env[61936]: DEBUG oslo_vmware.api [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.518688] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.519273] env[61936]: DEBUG nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 796.521725] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.003s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.523106] env[61936]: INFO nova.compute.claims [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.895997] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Releasing lock "refresh_cache-50cb3f79-39b2-46ab-a35c-fdad3bb266cd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.896383] env[61936]: DEBUG nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Instance network_info: |[{"id": "1056fca0-7eeb-464b-befc-9a19819b7a58", "address": "fa:16:3e:cf:5b:c7", "network": {"id": "341c94d2-35e4-44d5-b231-2e36201f4ef7", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "0a2bb4cb48fe49aba25dc72f9f3195d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1056fca0-7e", "ovs_interfaceid": "1056fca0-7eeb-464b-befc-9a19819b7a58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 796.896688] env[61936]: DEBUG oslo_concurrency.lockutils [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] Acquired lock "refresh_cache-50cb3f79-39b2-46ab-a35c-fdad3bb266cd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.896863] env[61936]: DEBUG nova.network.neutron [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Refreshing network info cache for port 1056fca0-7eeb-464b-befc-9a19819b7a58 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.898176] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:5b:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04e15990-16e1-4cb2-b0f0-06c362e68c5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1056fca0-7eeb-464b-befc-9a19819b7a58', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.907771] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Creating folder: Project (2b7ed656a83743ddb23b452c38d5b607). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.908969] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b401739-9bd8-4daf-a057-0df2be52eb8a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.923490] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Created folder: Project (2b7ed656a83743ddb23b452c38d5b607) in parent group-v269874. [ 796.923490] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Creating folder: Instances. Parent ref: group-v269895. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.923490] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-218bf2e0-4d8a-441b-a795-bf822ce4f7e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.932381] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Created folder: Instances in parent group-v269895. [ 796.932381] env[61936]: DEBUG oslo.service.loopingcall [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.932551] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 796.932753] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6bc16ff0-1c71-4d5c-93fa-07332c626b03 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.951685] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.951685] env[61936]: value = "task-1252790" [ 796.951685] env[61936]: _type = "Task" [ 796.951685] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.961016] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252790, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.019761] env[61936]: DEBUG oslo_vmware.api [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252787, 'name': PowerOffVM_Task, 'duration_secs': 0.208771} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.021193] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.021193] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.021193] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d397b4e-e3c6-4589-abc7-99fb26da08bb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.029669] env[61936]: DEBUG nova.compute.utils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.031935] env[61936]: DEBUG nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 797.032134] env[61936]: DEBUG nova.network.neutron [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 797.043024] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.043249] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.043420] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Deleting the datastore file [datastore1] 498a77d6-ca0c-4841-b4a2-b7a024281c6a {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.043659] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a77b09fc-10fd-4fa5-9617-16c178bf01f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.050184] env[61936]: DEBUG oslo_vmware.api [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for the task: (returnval){ [ 797.050184] env[61936]: value = "task-1252792" [ 797.050184] env[61936]: _type = "Task" [ 797.050184] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.059166] env[61936]: DEBUG oslo_vmware.api [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252792, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.072922] env[61936]: DEBUG nova.policy [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '471a0d8905b44e43bce9aa5b0945d55d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3c91fa7bf4742c283f34eff689c7502', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.370502] env[61936]: DEBUG nova.network.neutron [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Successfully created port: 87c99a35-e6ce-45e2-8354-80f5922cea14 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.451445] env[61936]: INFO nova.scheduler.client.report [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Deleted allocations for instance 0b14583c-dd0f-429d-a386-2acda353465a [ 797.472498] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252790, 'name': CreateVM_Task, 'duration_secs': 0.298858} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.472498] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.482941] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.484843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.484843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.484843] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-758ba46e-982a-47d3-97b5-8b3455561419 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.493415] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 797.493415] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e52843-7cc5-2d73-1434-bcdcca75037e" [ 797.493415] env[61936]: _type = "Task" [ 797.493415] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.506420] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e52843-7cc5-2d73-1434-bcdcca75037e, 'name': SearchDatastore_Task, 'duration_secs': 0.009071} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.509333] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.509333] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.509333] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.509333] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.509485] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.509485] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9df5bb1e-496d-48a5-924e-c87dfe69d07b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.515390] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.516148] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.516390] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-664cb599-6f32-452f-8109-f5dc6f4bfaff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.521151] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 797.521151] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5246c604-83af-fe7c-5413-62d42317c031" [ 797.521151] env[61936]: _type = "Task" [ 797.521151] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.529183] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5246c604-83af-fe7c-5413-62d42317c031, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.533189] env[61936]: DEBUG nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 797.560148] env[61936]: DEBUG oslo_vmware.api [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Task: {'id': task-1252792, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09995} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.560401] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.560581] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.560754] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.560923] env[61936]: INFO nova.compute.manager [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Took 1.07 seconds to destroy the instance on the hypervisor. [ 797.561209] env[61936]: DEBUG oslo.service.loopingcall [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.561648] env[61936]: DEBUG nova.compute.manager [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 797.561746] env[61936]: DEBUG nova.network.neutron [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.583653] env[61936]: DEBUG nova.network.neutron [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.902756] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcecef7-6712-4b26-8bb1-60cba6ab999a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.912922] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600d7411-435e-4b0f-9342-a82d5e022306 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.944935] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85b9ca2-260c-4a7e-892a-80098d82fa39 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.952617] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2e1ff8-baae-4a6c-ac5c-aea716518d35 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.967745] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a3941ef-943e-4f08-aee0-fc4be25cf4f3 tempest-ServersTestBootFromVolume-872234183 tempest-ServersTestBootFromVolume-872234183-project-member] Lock "0b14583c-dd0f-429d-a386-2acda353465a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.310s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.967745] env[61936]: DEBUG nova.compute.provider_tree [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.969196] env[61936]: DEBUG nova.network.neutron [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Updated VIF entry in instance network info cache for port 1056fca0-7eeb-464b-befc-9a19819b7a58. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 797.969493] env[61936]: DEBUG nova.network.neutron [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Updating instance_info_cache with network_info: [{"id": "1056fca0-7eeb-464b-befc-9a19819b7a58", "address": "fa:16:3e:cf:5b:c7", "network": {"id": "341c94d2-35e4-44d5-b231-2e36201f4ef7", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.159", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "0a2bb4cb48fe49aba25dc72f9f3195d6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1056fca0-7e", "ovs_interfaceid": "1056fca0-7eeb-464b-befc-9a19819b7a58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.038182] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5246c604-83af-fe7c-5413-62d42317c031, 'name': SearchDatastore_Task, 'duration_secs': 0.008261} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.039136] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea980c43-4832-4cfb-843e-0bff61a4b0ff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.048308] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 798.048308] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528d8700-54ad-aaae-c946-e7fc4e01b5ea" [ 798.048308] env[61936]: _type = "Task" [ 798.048308] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.055983] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528d8700-54ad-aaae-c946-e7fc4e01b5ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.088478] env[61936]: DEBUG nova.network.neutron [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.473216] env[61936]: DEBUG nova.scheduler.client.report [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 798.477121] env[61936]: DEBUG oslo_concurrency.lockutils [req-900bcbe1-5a48-4d16-a5e1-d4b73d6a0fe1 req-de2dbbc9-7b31-42c8-bf27-38ab11480efc service nova] Releasing lock "refresh_cache-50cb3f79-39b2-46ab-a35c-fdad3bb266cd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.483685] env[61936]: DEBUG nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 798.547558] env[61936]: DEBUG nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 798.560151] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528d8700-54ad-aaae-c946-e7fc4e01b5ea, 'name': SearchDatastore_Task, 'duration_secs': 0.008659} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.560425] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.560685] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 50cb3f79-39b2-46ab-a35c-fdad3bb266cd/50cb3f79-39b2-46ab-a35c-fdad3bb266cd.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 798.560947] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f730c70-e4ff-4559-b391-cd9262fd0806 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.569434] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 798.569434] env[61936]: value = "task-1252793" [ 798.569434] env[61936]: _type = "Task" [ 798.569434] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.574832] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 798.575100] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.575270] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 798.575450] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.575591] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 798.575731] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 798.575974] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 798.576159] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 798.576326] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 798.576484] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 798.576651] env[61936]: DEBUG nova.virt.hardware [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 798.577424] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3066b1da-e181-404f-8d88-599f9d682066 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.582606] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252793, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.587721] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2ec793-1c56-4852-a5ca-ef3d7a6ab700 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.591461] env[61936]: INFO nova.compute.manager [-] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Took 1.03 seconds to deallocate network for instance. [ 798.978743] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.979455] env[61936]: DEBUG nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 798.983417] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.923s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.985307] env[61936]: INFO nova.compute.claims [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.015247] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.078319] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252793, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50298} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.078609] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 50cb3f79-39b2-46ab-a35c-fdad3bb266cd/50cb3f79-39b2-46ab-a35c-fdad3bb266cd.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.078783] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.079061] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-480e3159-762c-4202-960a-c84f59d43615 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.085136] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 799.085136] env[61936]: value = "task-1252795" [ 799.085136] env[61936]: _type = "Task" [ 799.085136] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.094340] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252795, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.106036] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.336197] env[61936]: DEBUG nova.compute.manager [req-7edb9c78-423d-4ef3-a1fa-b988f3dd7a33 req-406dbf99-5325-4fd0-ab41-bc055b19bcce service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Received event network-vif-plugged-87c99a35-e6ce-45e2-8354-80f5922cea14 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 799.336197] env[61936]: DEBUG oslo_concurrency.lockutils [req-7edb9c78-423d-4ef3-a1fa-b988f3dd7a33 req-406dbf99-5325-4fd0-ab41-bc055b19bcce service nova] Acquiring lock "1251e7f5-684c-428f-9d63-60b77084d09e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.336197] env[61936]: DEBUG oslo_concurrency.lockutils [req-7edb9c78-423d-4ef3-a1fa-b988f3dd7a33 req-406dbf99-5325-4fd0-ab41-bc055b19bcce service nova] Lock "1251e7f5-684c-428f-9d63-60b77084d09e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.336197] env[61936]: DEBUG oslo_concurrency.lockutils [req-7edb9c78-423d-4ef3-a1fa-b988f3dd7a33 req-406dbf99-5325-4fd0-ab41-bc055b19bcce service nova] Lock "1251e7f5-684c-428f-9d63-60b77084d09e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.336197] env[61936]: DEBUG nova.compute.manager [req-7edb9c78-423d-4ef3-a1fa-b988f3dd7a33 req-406dbf99-5325-4fd0-ab41-bc055b19bcce service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] No waiting events found dispatching network-vif-plugged-87c99a35-e6ce-45e2-8354-80f5922cea14 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 799.336414] env[61936]: WARNING nova.compute.manager [req-7edb9c78-423d-4ef3-a1fa-b988f3dd7a33 req-406dbf99-5325-4fd0-ab41-bc055b19bcce service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Received unexpected event network-vif-plugged-87c99a35-e6ce-45e2-8354-80f5922cea14 for instance with vm_state building and task_state spawning. [ 799.433957] env[61936]: DEBUG nova.network.neutron [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Successfully updated port: 87c99a35-e6ce-45e2-8354-80f5922cea14 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.491162] env[61936]: DEBUG nova.compute.utils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.492513] env[61936]: DEBUG nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 799.492682] env[61936]: DEBUG nova.network.neutron [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.542285] env[61936]: DEBUG nova.policy [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcee2be71ae044338df4422a3c575ca4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e81b119251964bdb8e5a0ea84b29f2a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.595635] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252795, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063961} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.595923] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.596711] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb800e7-5476-4016-bc6e-fcf048c9b728 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.619153] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 50cb3f79-39b2-46ab-a35c-fdad3bb266cd/50cb3f79-39b2-46ab-a35c-fdad3bb266cd.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.619388] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa4c953f-ec44-41f9-92de-8b21aad44aad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.640448] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 799.640448] env[61936]: value = "task-1252796" [ 799.640448] env[61936]: _type = "Task" [ 799.640448] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.648127] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252796, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.846975] env[61936]: DEBUG nova.network.neutron [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Successfully created port: 670384d2-16a5-4c63-a597-a1513479b68b {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.936049] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "refresh_cache-1251e7f5-684c-428f-9d63-60b77084d09e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.936049] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquired lock "refresh_cache-1251e7f5-684c-428f-9d63-60b77084d09e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.936049] env[61936]: DEBUG nova.network.neutron [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.003161] env[61936]: DEBUG nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 800.154853] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252796, 'name': ReconfigVM_Task, 'duration_secs': 0.314421} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.155231] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 50cb3f79-39b2-46ab-a35c-fdad3bb266cd/50cb3f79-39b2-46ab-a35c-fdad3bb266cd.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.156043] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c97371fe-1237-4038-8a9c-c6a0a3ba2e4f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.167485] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 800.167485] env[61936]: value = "task-1252797" [ 800.167485] env[61936]: _type = "Task" [ 800.167485] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.176179] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252797, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.393067] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c523708b-51be-47e2-a632-1e910e4ed58a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.401124] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ce4fdf-8961-4bce-bee3-8745169b1f86 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.435309] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517fe57e-6f4d-4253-a995-ecd6eb5a89d0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.445037] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef85e1e-e475-4e7c-a73a-a5c3914d4e54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.461666] env[61936]: DEBUG nova.compute.provider_tree [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.468102] env[61936]: DEBUG nova.network.neutron [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.629259] env[61936]: DEBUG nova.network.neutron [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Updating instance_info_cache with network_info: [{"id": "87c99a35-e6ce-45e2-8354-80f5922cea14", "address": "fa:16:3e:29:ad:eb", "network": {"id": "9581352a-900e-451c-be26-9b986e1781ce", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1981629710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c91fa7bf4742c283f34eff689c7502", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87c99a35-e6", "ovs_interfaceid": "87c99a35-e6ce-45e2-8354-80f5922cea14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.677255] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252797, 'name': Rename_Task, 'duration_secs': 0.156811} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.677508] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.677913] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a87720b-933c-4822-8fc9-fca252ee2c84 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.685630] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 800.685630] env[61936]: value = "task-1252798" [ 800.685630] env[61936]: _type = "Task" [ 800.685630] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.692705] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.964693] env[61936]: DEBUG nova.scheduler.client.report [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 801.014972] env[61936]: DEBUG nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 801.041630] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 801.041910] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.042092] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.042319] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.042488] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.042667] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 801.042914] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 801.043183] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 801.043459] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 801.043725] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 801.043953] env[61936]: DEBUG nova.virt.hardware [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 801.045370] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a299bf8-8a10-430e-9d06-1e17467b85c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.053992] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95a682b-94d1-488a-81de-87eb0aa42a31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.132128] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Releasing lock "refresh_cache-1251e7f5-684c-428f-9d63-60b77084d09e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.132479] env[61936]: DEBUG nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Instance network_info: |[{"id": "87c99a35-e6ce-45e2-8354-80f5922cea14", "address": "fa:16:3e:29:ad:eb", "network": {"id": "9581352a-900e-451c-be26-9b986e1781ce", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1981629710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c91fa7bf4742c283f34eff689c7502", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87c99a35-e6", "ovs_interfaceid": "87c99a35-e6ce-45e2-8354-80f5922cea14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 801.132900] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:ad:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5446413d-c3b0-4cd2-a962-62240db178ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87c99a35-e6ce-45e2-8354-80f5922cea14', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.140725] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Creating folder: Project (a3c91fa7bf4742c283f34eff689c7502). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.141277] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-254d9db3-06ca-4e00-9157-fcb7fd838883 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.151849] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Created folder: Project (a3c91fa7bf4742c283f34eff689c7502) in parent group-v269874. [ 801.152060] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Creating folder: Instances. Parent ref: group-v269898. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.152308] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ec95c13-4f6d-4083-a9fa-b60f9f464aea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.161335] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Created folder: Instances in parent group-v269898. [ 801.161581] env[61936]: DEBUG oslo.service.loopingcall [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.161756] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 801.162511] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c82d56da-3e74-4004-94f6-f9bd4edc94ff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.181516] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.181516] env[61936]: value = "task-1252801" [ 801.181516] env[61936]: _type = "Task" [ 801.181516] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.192815] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252801, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.198023] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252798, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.413417] env[61936]: DEBUG nova.compute.manager [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Received event network-changed-87c99a35-e6ce-45e2-8354-80f5922cea14 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 801.413417] env[61936]: DEBUG nova.compute.manager [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Refreshing instance network info cache due to event network-changed-87c99a35-e6ce-45e2-8354-80f5922cea14. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 801.413417] env[61936]: DEBUG oslo_concurrency.lockutils [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] Acquiring lock "refresh_cache-1251e7f5-684c-428f-9d63-60b77084d09e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.414529] env[61936]: DEBUG oslo_concurrency.lockutils [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] Acquired lock "refresh_cache-1251e7f5-684c-428f-9d63-60b77084d09e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.414529] env[61936]: DEBUG nova.network.neutron [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Refreshing network info cache for port 87c99a35-e6ce-45e2-8354-80f5922cea14 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.470518] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.471172] env[61936]: DEBUG nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 801.473978] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.175s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.474132] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.474715] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 801.474715] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.451s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.477266] env[61936]: INFO nova.compute.claims [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.482127] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208d7418-8c2e-4345-ac86-ec7fbbfcc912 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.490372] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fbb978-f1bd-4a6d-88c5-fc45539a5364 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.505228] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e70f169-dbdb-409e-bcbd-ce1c2f107e91 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.511900] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a202c5-bced-4867-8192-61297bc7e427 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.541747] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181532MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 801.541899] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.680091] env[61936]: DEBUG nova.network.neutron [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Successfully updated port: 670384d2-16a5-4c63-a597-a1513479b68b {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.693832] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252801, 'name': CreateVM_Task, 'duration_secs': 0.352992} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.696626] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.697211] env[61936]: DEBUG oslo_vmware.api [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252798, 'name': PowerOnVM_Task, 'duration_secs': 0.516725} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.697841] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.698029] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.698336] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.698594] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.698783] env[61936]: INFO nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Took 7.50 seconds to spawn the instance on the hypervisor. [ 801.698971] env[61936]: DEBUG nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 801.699236] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aac5c679-5752-4d61-b22c-f4cfdf971c6c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.701259] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d7ae1c-83b6-4bd4-b261-d56c870f83de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.708300] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 801.708300] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52adca92-5fea-10f3-d624-9986fc7e6709" [ 801.708300] env[61936]: _type = "Task" [ 801.708300] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.720218] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52adca92-5fea-10f3-d624-9986fc7e6709, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.982154] env[61936]: DEBUG nova.compute.utils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.983615] env[61936]: DEBUG nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 801.983845] env[61936]: DEBUG nova.network.neutron [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 802.073376] env[61936]: DEBUG nova.policy [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e983f2db914404a93cd8d9e80c2a0cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f29380d7b4344718d2a6dd238b843a2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.185782] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-144ddda8-db07-4308-8fd3-4045067b5b37" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.185782] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-144ddda8-db07-4308-8fd3-4045067b5b37" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.185782] env[61936]: DEBUG nova.network.neutron [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.222496] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52adca92-5fea-10f3-d624-9986fc7e6709, 'name': SearchDatastore_Task, 'duration_secs': 0.032541} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.223196] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.226637] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.226637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.226637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.226637] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.226637] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9d73e72-ca93-4f71-a5da-7d01a238f404 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.228524] env[61936]: INFO nova.compute.manager [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Took 26.62 seconds to build instance. [ 802.234214] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.234407] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.235157] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e560b65-1850-4df7-adef-d7dfab19e64a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.240770] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 802.240770] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5299a9f6-f2b4-3b26-61ce-45ee8f988b05" [ 802.240770] env[61936]: _type = "Task" [ 802.240770] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.249654] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5299a9f6-f2b4-3b26-61ce-45ee8f988b05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.487549] env[61936]: DEBUG nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 802.544660] env[61936]: DEBUG nova.network.neutron [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Updated VIF entry in instance network info cache for port 87c99a35-e6ce-45e2-8354-80f5922cea14. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 802.548877] env[61936]: DEBUG nova.network.neutron [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Updating instance_info_cache with network_info: [{"id": "87c99a35-e6ce-45e2-8354-80f5922cea14", "address": "fa:16:3e:29:ad:eb", "network": {"id": "9581352a-900e-451c-be26-9b986e1781ce", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1981629710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3c91fa7bf4742c283f34eff689c7502", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5446413d-c3b0-4cd2-a962-62240db178ac", "external-id": "nsx-vlan-transportzone-528", "segmentation_id": 528, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87c99a35-e6", "ovs_interfaceid": "87c99a35-e6ce-45e2-8354-80f5922cea14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.732543] env[61936]: DEBUG oslo_concurrency.lockutils [None req-83a79283-3899-4c80-a711-d36569956a7e tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.379s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.755751] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5299a9f6-f2b4-3b26-61ce-45ee8f988b05, 'name': SearchDatastore_Task, 'duration_secs': 0.009109} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.756953] env[61936]: DEBUG nova.network.neutron [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.759656] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4991b555-f8b5-4eaa-a769-91dee8bf7335 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.767322] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 802.767322] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523a22e5-b512-f57d-f011-9cb4714b2044" [ 802.767322] env[61936]: _type = "Task" [ 802.767322] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.775238] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523a22e5-b512-f57d-f011-9cb4714b2044, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.841305] env[61936]: DEBUG nova.network.neutron [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Successfully created port: c319d6d8-72f1-4705-8ee6-1603b01cf344 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.864918] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22420f90-eac2-4d59-868b-14848b3fb16e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.873313] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdb0c81-4e22-4d4f-b6df-b748819740be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.905132] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ddda6c-2751-4d07-9d89-3bf0a1992425 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.912664] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684d0daa-373d-4355-a2e3-382d8c216793 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.931089] env[61936]: DEBUG nova.compute.provider_tree [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.055476] env[61936]: DEBUG oslo_concurrency.lockutils [req-45ca2309-aa94-40ae-b687-d42218c70ed2 req-2fc40701-e8c7-417f-a224-1608b22f2564 service nova] Releasing lock "refresh_cache-1251e7f5-684c-428f-9d63-60b77084d09e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.092966] env[61936]: DEBUG nova.network.neutron [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Updating instance_info_cache with network_info: [{"id": "670384d2-16a5-4c63-a597-a1513479b68b", "address": "fa:16:3e:f8:c0:35", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670384d2-16", "ovs_interfaceid": "670384d2-16a5-4c63-a597-a1513479b68b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.237081] env[61936]: DEBUG nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 803.278281] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523a22e5-b512-f57d-f011-9cb4714b2044, 'name': SearchDatastore_Task, 'duration_secs': 0.009142} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.278634] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.279231] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1251e7f5-684c-428f-9d63-60b77084d09e/1251e7f5-684c-428f-9d63-60b77084d09e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.279648] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-911ed935-cd0d-43ed-8620-28f1bbe56bbe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.289438] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 803.289438] env[61936]: value = "task-1252802" [ 803.289438] env[61936]: _type = "Task" [ 803.289438] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.296553] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252802, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.434556] env[61936]: DEBUG nova.scheduler.client.report [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 803.441896] env[61936]: DEBUG nova.compute.manager [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Received event network-vif-plugged-670384d2-16a5-4c63-a597-a1513479b68b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 803.442025] env[61936]: DEBUG oslo_concurrency.lockutils [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] Acquiring lock "144ddda8-db07-4308-8fd3-4045067b5b37-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.442241] env[61936]: DEBUG oslo_concurrency.lockutils [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] Lock "144ddda8-db07-4308-8fd3-4045067b5b37-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.442455] env[61936]: DEBUG oslo_concurrency.lockutils [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] Lock "144ddda8-db07-4308-8fd3-4045067b5b37-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.442601] env[61936]: DEBUG nova.compute.manager [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] No waiting events found dispatching network-vif-plugged-670384d2-16a5-4c63-a597-a1513479b68b {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 803.442792] env[61936]: WARNING nova.compute.manager [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Received unexpected event network-vif-plugged-670384d2-16a5-4c63-a597-a1513479b68b for instance with vm_state building and task_state spawning. [ 803.442947] env[61936]: DEBUG nova.compute.manager [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Received event network-changed-670384d2-16a5-4c63-a597-a1513479b68b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 803.443117] env[61936]: DEBUG nova.compute.manager [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Refreshing instance network info cache due to event network-changed-670384d2-16a5-4c63-a597-a1513479b68b. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 803.443335] env[61936]: DEBUG oslo_concurrency.lockutils [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] Acquiring lock "refresh_cache-144ddda8-db07-4308-8fd3-4045067b5b37" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.499322] env[61936]: DEBUG nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 803.527805] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 803.528099] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.528289] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 803.528451] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.528592] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 803.528737] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 803.528964] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 803.529181] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 803.529355] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 803.529517] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 803.529683] env[61936]: DEBUG nova.virt.hardware [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 803.530634] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e90cdb1-dcb6-421d-9aca-d576eca1cbff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.538587] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cecd4bf-faae-4568-9b64-a8d57249ee46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.595836] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-144ddda8-db07-4308-8fd3-4045067b5b37" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.596285] env[61936]: DEBUG nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Instance network_info: |[{"id": "670384d2-16a5-4c63-a597-a1513479b68b", "address": "fa:16:3e:f8:c0:35", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670384d2-16", "ovs_interfaceid": "670384d2-16a5-4c63-a597-a1513479b68b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 803.596939] env[61936]: DEBUG oslo_concurrency.lockutils [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] Acquired lock "refresh_cache-144ddda8-db07-4308-8fd3-4045067b5b37" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.596939] env[61936]: DEBUG nova.network.neutron [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Refreshing network info cache for port 670384d2-16a5-4c63-a597-a1513479b68b {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 803.598363] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:c0:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '670384d2-16a5-4c63-a597-a1513479b68b', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.605855] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating folder: Project (e81b119251964bdb8e5a0ea84b29f2a1). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.606466] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0169d23f-f3fa-449c-895a-2f139542cc26 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.617077] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created folder: Project (e81b119251964bdb8e5a0ea84b29f2a1) in parent group-v269874. [ 803.617254] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating folder: Instances. Parent ref: group-v269901. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.617534] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5fea8142-8eb2-4cf4-af1a-ca4a15d2bd3c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.627112] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created folder: Instances in parent group-v269901. [ 803.627112] env[61936]: DEBUG oslo.service.loopingcall [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.627286] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 803.628034] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8adbc1dc-3589-4c2d-8997-2e041273283c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.646240] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.646240] env[61936]: value = "task-1252805" [ 803.646240] env[61936]: _type = "Task" [ 803.646240] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.654878] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252805, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.765315] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.799760] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252802, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46341} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.800073] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1251e7f5-684c-428f-9d63-60b77084d09e/1251e7f5-684c-428f-9d63-60b77084d09e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.800295] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.800554] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0996eda5-2283-4b3e-9e05-516916605383 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.809842] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 803.809842] env[61936]: value = "task-1252806" [ 803.809842] env[61936]: _type = "Task" [ 803.809842] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.817669] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252806, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.943316] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.943316] env[61936]: DEBUG nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 803.945427] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.893s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.156662] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252805, 'name': CreateVM_Task, 'duration_secs': 0.356567} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.158910] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.159639] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.159877] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.160261] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.160812] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c28540f-4c90-4548-8ce7-84bfc94062d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.165627] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 804.165627] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b463d0-f4f9-e51e-f8e5-46886ccab4bc" [ 804.165627] env[61936]: _type = "Task" [ 804.165627] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.173387] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b463d0-f4f9-e51e-f8e5-46886ccab4bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.318843] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252806, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086474} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.319209] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.319989] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a919a80-428b-4fa5-85cc-29fb658e9635 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.341565] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 1251e7f5-684c-428f-9d63-60b77084d09e/1251e7f5-684c-428f-9d63-60b77084d09e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.343990] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bde0be50-3826-4f90-bda4-ae6ee027dd53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.362850] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 804.362850] env[61936]: value = "task-1252807" [ 804.362850] env[61936]: _type = "Task" [ 804.362850] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.371535] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252807, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.447267] env[61936]: DEBUG nova.compute.utils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.448937] env[61936]: DEBUG nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 804.452920] env[61936]: DEBUG nova.network.neutron [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 804.464117] env[61936]: DEBUG nova.network.neutron [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Updated VIF entry in instance network info cache for port 670384d2-16a5-4c63-a597-a1513479b68b. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.464117] env[61936]: DEBUG nova.network.neutron [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Updating instance_info_cache with network_info: [{"id": "670384d2-16a5-4c63-a597-a1513479b68b", "address": "fa:16:3e:f8:c0:35", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap670384d2-16", "ovs_interfaceid": "670384d2-16a5-4c63-a597-a1513479b68b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.531482] env[61936]: DEBUG nova.policy [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '614efde420c149bb898c59875c71c060', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b8b985036b44abd9e833f7b99c0e27f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 804.677329] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b463d0-f4f9-e51e-f8e5-46886ccab4bc, 'name': SearchDatastore_Task, 'duration_secs': 0.010392} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.680475] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.680616] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.680836] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.680969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.681166] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.681793] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-585c121f-59cc-4eb6-b40e-eede83629a55 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.692441] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.692441] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.692441] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c18602ec-919b-43b6-b7b0-7ce383986a9c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.695867] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 804.695867] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523eb8bf-283c-4279-f9bc-f2221828f482" [ 804.695867] env[61936]: _type = "Task" [ 804.695867] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.704848] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523eb8bf-283c-4279-f9bc-f2221828f482, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.772163] env[61936]: DEBUG nova.network.neutron [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Successfully updated port: c319d6d8-72f1-4705-8ee6-1603b01cf344 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.869852] env[61936]: DEBUG nova.network.neutron [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Successfully created port: fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.872334] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecccb8f5-aa54-4f8e-881f-8a600415b53f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.880621] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.882980] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9f3227-fbe7-48df-9867-8e1bb3476197 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.916754] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8b9ab5-7862-4d7e-b158-cef1a36df1b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.924991] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e58aa1a-372b-4092-83d4-940819ceb2dc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.942203] env[61936]: DEBUG nova.compute.provider_tree [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.956690] env[61936]: DEBUG nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 804.972974] env[61936]: DEBUG oslo_concurrency.lockutils [req-eef76912-ca84-42da-9f7d-f15ad065d1c2 req-a9246a0c-020d-4d6e-b929-08114f495927 service nova] Releasing lock "refresh_cache-144ddda8-db07-4308-8fd3-4045067b5b37" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.206803] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523eb8bf-283c-4279-f9bc-f2221828f482, 'name': SearchDatastore_Task, 'duration_secs': 0.008899} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.207376] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b0fc816-8ea5-4351-b097-1d6aebdfbb86 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.212805] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 805.212805] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b1ebca-915d-f3c9-b547-a9eb4961dcfd" [ 805.212805] env[61936]: _type = "Task" [ 805.212805] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.220385] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b1ebca-915d-f3c9-b547-a9eb4961dcfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.274049] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "refresh_cache-867a4bdc-8326-48b3-ba17-ca8336533c2d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.274208] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquired lock "refresh_cache-867a4bdc-8326-48b3-ba17-ca8336533c2d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.274372] env[61936]: DEBUG nova.network.neutron [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.372966] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252807, 'name': ReconfigVM_Task, 'duration_secs': 0.600833} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.373259] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 1251e7f5-684c-428f-9d63-60b77084d09e/1251e7f5-684c-428f-9d63-60b77084d09e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.373861] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09292063-76a4-4ca9-a233-4e42e29fc13f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.379610] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 805.379610] env[61936]: value = "task-1252808" [ 805.379610] env[61936]: _type = "Task" [ 805.379610] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.387385] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252808, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.445590] env[61936]: DEBUG nova.scheduler.client.report [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 805.469733] env[61936]: DEBUG nova.compute.manager [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Received event network-vif-plugged-c319d6d8-72f1-4705-8ee6-1603b01cf344 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 805.469966] env[61936]: DEBUG oslo_concurrency.lockutils [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] Acquiring lock "867a4bdc-8326-48b3-ba17-ca8336533c2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.470226] env[61936]: DEBUG oslo_concurrency.lockutils [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.470382] env[61936]: DEBUG oslo_concurrency.lockutils [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.470548] env[61936]: DEBUG nova.compute.manager [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] No waiting events found dispatching network-vif-plugged-c319d6d8-72f1-4705-8ee6-1603b01cf344 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 805.470710] env[61936]: WARNING nova.compute.manager [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Received unexpected event network-vif-plugged-c319d6d8-72f1-4705-8ee6-1603b01cf344 for instance with vm_state building and task_state spawning. [ 805.470824] env[61936]: DEBUG nova.compute.manager [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Received event network-changed-c319d6d8-72f1-4705-8ee6-1603b01cf344 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 805.470961] env[61936]: DEBUG nova.compute.manager [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Refreshing instance network info cache due to event network-changed-c319d6d8-72f1-4705-8ee6-1603b01cf344. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 805.471261] env[61936]: DEBUG oslo_concurrency.lockutils [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] Acquiring lock "refresh_cache-867a4bdc-8326-48b3-ba17-ca8336533c2d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.723016] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b1ebca-915d-f3c9-b547-a9eb4961dcfd, 'name': SearchDatastore_Task, 'duration_secs': 0.010825} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.723291] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.723545] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 144ddda8-db07-4308-8fd3-4045067b5b37/144ddda8-db07-4308-8fd3-4045067b5b37.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.723790] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3517dffd-c7f8-4efc-a61c-caa666961b4b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.730020] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 805.730020] env[61936]: value = "task-1252809" [ 805.730020] env[61936]: _type = "Task" [ 805.730020] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.737940] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252809, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.816350] env[61936]: DEBUG nova.network.neutron [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.892044] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252808, 'name': Rename_Task, 'duration_secs': 0.138384} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.892044] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.892044] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98129057-02b9-4c2d-bc9b-453ca4aed910 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.899186] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 805.899186] env[61936]: value = "task-1252810" [ 805.899186] env[61936]: _type = "Task" [ 805.899186] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.906327] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252810, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.954022] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.006s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.954022] env[61936]: ERROR nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. [ 805.954022] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Traceback (most recent call last): [ 805.954022] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 805.954022] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self.driver.spawn(context, instance, image_meta, [ 805.954022] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 805.954022] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self._vmops.spawn(context, instance, image_meta, injected_files, [ 805.954022] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 805.954022] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] vm_ref = self.build_virtual_machine(instance, [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] vif_infos = vmwarevif.get_vif_info(self._session, [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] for vif in network_info: [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return self._sync_wrapper(fn, *args, **kwargs) [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self.wait() [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self[:] = self._gt.wait() [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return self._exit_event.wait() [ 805.954459] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] result = hub.switch() [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return self.greenlet.switch() [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] result = function(*args, **kwargs) [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] return func(*args, **kwargs) [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] raise e [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] nwinfo = self.network_api.allocate_for_instance( [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 805.954815] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] created_port_ids = self._update_ports_for_instance( [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] with excutils.save_and_reraise_exception(): [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] self.force_reraise() [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] raise self.value [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] updated_port = self._update_port( [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] _ensure_no_port_binding_failure(port) [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.955195] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] raise exception.PortBindingFailed(port_id=port['id']) [ 805.955542] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] nova.exception.PortBindingFailed: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. [ 805.955542] env[61936]: ERROR nova.compute.manager [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] [ 805.955542] env[61936]: DEBUG nova.compute.utils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 805.961384] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Build of instance 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85 was re-scheduled: Binding failed for port 781d1814-e806-4aeb-87d9-ed6a427f2194, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 805.961384] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 805.961384] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquiring lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.961384] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Acquired lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.961782] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.961782] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.684s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.969952] env[61936]: DEBUG nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 806.011641] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 806.011879] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.012270] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 806.012509] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.012677] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 806.012835] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 806.013060] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 806.013220] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 806.013381] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 806.013538] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 806.013705] env[61936]: DEBUG nova.virt.hardware [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 806.014970] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b28b72-62fc-45de-85b7-ef5630ce24a1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.024489] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a8fd11-0a9a-4d46-abe6-24fb7cfc18c5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.077625] env[61936]: DEBUG nova.network.neutron [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Updating instance_info_cache with network_info: [{"id": "c319d6d8-72f1-4705-8ee6-1603b01cf344", "address": "fa:16:3e:ee:08:5a", "network": {"id": "718536d2-ce57-4e71-acd9-205d609bca9d", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1353227355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f29380d7b4344718d2a6dd238b843a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc319d6d8-72", "ovs_interfaceid": "c319d6d8-72f1-4705-8ee6-1603b01cf344", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.239641] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252809, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435953} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.239932] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 144ddda8-db07-4308-8fd3-4045067b5b37/144ddda8-db07-4308-8fd3-4045067b5b37.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 806.240680] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.240680] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b101ac33-a660-4a0d-a6e0-55bc9422865d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.246936] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 806.246936] env[61936]: value = "task-1252811" [ 806.246936] env[61936]: _type = "Task" [ 806.246936] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.254746] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.322029] env[61936]: DEBUG nova.compute.manager [req-0910fbb6-8e94-4145-80b0-39f185cc612c req-392d7ce3-15a2-49db-bf06-e4250c1ee323 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Received event network-vif-plugged-fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 806.322029] env[61936]: DEBUG oslo_concurrency.lockutils [req-0910fbb6-8e94-4145-80b0-39f185cc612c req-392d7ce3-15a2-49db-bf06-e4250c1ee323 service nova] Acquiring lock "301b0504-5cf9-44e0-bd3e-342f8a89278b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.322029] env[61936]: DEBUG oslo_concurrency.lockutils [req-0910fbb6-8e94-4145-80b0-39f185cc612c req-392d7ce3-15a2-49db-bf06-e4250c1ee323 service nova] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.322029] env[61936]: DEBUG oslo_concurrency.lockutils [req-0910fbb6-8e94-4145-80b0-39f185cc612c req-392d7ce3-15a2-49db-bf06-e4250c1ee323 service nova] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.322029] env[61936]: DEBUG nova.compute.manager [req-0910fbb6-8e94-4145-80b0-39f185cc612c req-392d7ce3-15a2-49db-bf06-e4250c1ee323 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] No waiting events found dispatching network-vif-plugged-fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 806.322282] env[61936]: WARNING nova.compute.manager [req-0910fbb6-8e94-4145-80b0-39f185cc612c req-392d7ce3-15a2-49db-bf06-e4250c1ee323 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Received unexpected event network-vif-plugged-fc749785-35a5-468f-b3a3-b28dd3392c1c for instance with vm_state building and task_state spawning. [ 806.406965] env[61936]: DEBUG oslo_vmware.api [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252810, 'name': PowerOnVM_Task, 'duration_secs': 0.477034} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.410986] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.410986] env[61936]: INFO nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Took 7.86 seconds to spawn the instance on the hypervisor. [ 806.410986] env[61936]: DEBUG nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 806.410986] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c00d87-7c69-4f45-907c-be33c8f9f74d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.494864] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.577620] env[61936]: DEBUG nova.network.neutron [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Successfully updated port: fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.580840] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Releasing lock "refresh_cache-867a4bdc-8326-48b3-ba17-ca8336533c2d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.580952] env[61936]: DEBUG nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Instance network_info: |[{"id": "c319d6d8-72f1-4705-8ee6-1603b01cf344", "address": "fa:16:3e:ee:08:5a", "network": {"id": "718536d2-ce57-4e71-acd9-205d609bca9d", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1353227355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f29380d7b4344718d2a6dd238b843a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc319d6d8-72", "ovs_interfaceid": "c319d6d8-72f1-4705-8ee6-1603b01cf344", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 806.581855] env[61936]: DEBUG oslo_concurrency.lockutils [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] Acquired lock "refresh_cache-867a4bdc-8326-48b3-ba17-ca8336533c2d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.583590] env[61936]: DEBUG nova.network.neutron [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Refreshing network info cache for port c319d6d8-72f1-4705-8ee6-1603b01cf344 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.585035] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:08:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8abee039-d93e-48a7-8911-6416a3e1ff30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c319d6d8-72f1-4705-8ee6-1603b01cf344', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.593158] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Creating folder: Project (4f29380d7b4344718d2a6dd238b843a2). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.597011] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2029d2e4-fd3c-441c-95df-3cccc86429bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.601571] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.611668] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Created folder: Project (4f29380d7b4344718d2a6dd238b843a2) in parent group-v269874. [ 806.611857] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Creating folder: Instances. Parent ref: group-v269904. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.612591] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-51364e65-208d-4aa3-b031-23cf4e4e37c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.624672] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Created folder: Instances in parent group-v269904. [ 806.624855] env[61936]: DEBUG oslo.service.loopingcall [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.625049] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 806.625258] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34a6ca0e-8640-4808-8f1c-773e16813479 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.648782] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.648782] env[61936]: value = "task-1252814" [ 806.648782] env[61936]: _type = "Task" [ 806.648782] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.655025] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252814, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.760590] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09965} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.761235] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.761756] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8a0a62-0c27-482e-a692-32c7817ee72d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.785788] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 144ddda8-db07-4308-8fd3-4045067b5b37/144ddda8-db07-4308-8fd3-4045067b5b37.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.788457] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-885b7f3c-b85d-427d-819f-3f0931cbdf62 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.809976] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 806.809976] env[61936]: value = "task-1252815" [ 806.809976] env[61936]: _type = "Task" [ 806.809976] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.820480] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252815, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.875613] env[61936]: DEBUG nova.network.neutron [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Updated VIF entry in instance network info cache for port c319d6d8-72f1-4705-8ee6-1603b01cf344. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.876065] env[61936]: DEBUG nova.network.neutron [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Updating instance_info_cache with network_info: [{"id": "c319d6d8-72f1-4705-8ee6-1603b01cf344", "address": "fa:16:3e:ee:08:5a", "network": {"id": "718536d2-ce57-4e71-acd9-205d609bca9d", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1353227355-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4f29380d7b4344718d2a6dd238b843a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc319d6d8-72", "ovs_interfaceid": "c319d6d8-72f1-4705-8ee6-1603b01cf344", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.901124] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def2a3c4-f814-4648-a065-21598e6059d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.909090] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27339c8-79cb-4f2b-b03d-9d0046b7cec4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.943962] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2feca2c-0a27-49ee-9690-711f8af3432d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.948772] env[61936]: INFO nova.compute.manager [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Took 29.43 seconds to build instance. [ 806.953049] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cdfffb-62dc-48e2-8e1e-a6395717def3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.966628] env[61936]: DEBUG nova.compute.provider_tree [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.083044] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.083248] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquired lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.083409] env[61936]: DEBUG nova.network.neutron [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.104600] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Releasing lock "refresh_cache-7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.104894] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 807.105034] env[61936]: DEBUG nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 807.105225] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 807.123297] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.159155] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252814, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.321565] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252815, 'name': ReconfigVM_Task, 'duration_secs': 0.291887} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.321847] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 144ddda8-db07-4308-8fd3-4045067b5b37/144ddda8-db07-4308-8fd3-4045067b5b37.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.322485] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62a11bd2-7bb5-4a04-aa53-c8c03ea8c9c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.328701] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 807.328701] env[61936]: value = "task-1252816" [ 807.328701] env[61936]: _type = "Task" [ 807.328701] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.336199] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252816, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.380930] env[61936]: DEBUG oslo_concurrency.lockutils [req-15be4504-7c31-4719-ab5a-fd49d094b517 req-06d10486-9f1a-4ba0-bfa2-d67eacc09cb9 service nova] Releasing lock "refresh_cache-867a4bdc-8326-48b3-ba17-ca8336533c2d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.451457] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4c7555cb-3922-4ba0-8191-a666d0c4eb32 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "1251e7f5-684c-428f-9d63-60b77084d09e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.603s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.469950] env[61936]: DEBUG nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 807.614935] env[61936]: DEBUG nova.network.neutron [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.626687] env[61936]: DEBUG nova.network.neutron [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.660899] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252814, 'name': CreateVM_Task, 'duration_secs': 0.51688} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.661084] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.661809] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.662248] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.662321] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.662849] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06df33a9-0506-4024-ba76-4d83c0ca4dbe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.666903] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 807.666903] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526c53d5-614a-b1b5-fc4d-6120483bcbfc" [ 807.666903] env[61936]: _type = "Task" [ 807.666903] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.674760] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526c53d5-614a-b1b5-fc4d-6120483bcbfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.757877] env[61936]: DEBUG nova.network.neutron [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Updating instance_info_cache with network_info: [{"id": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "address": "fa:16:3e:c2:65:76", "network": {"id": "6a8ce523-f84f-45dc-b062-e340c36e0dc2", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1341537324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b8b985036b44abd9e833f7b99c0e27f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc749785-35", "ovs_interfaceid": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.838993] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252816, 'name': Rename_Task, 'duration_secs': 0.134988} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.839293] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 807.839530] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-315e6851-7b8e-4193-936c-dc331271bf7b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.845709] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 807.845709] env[61936]: value = "task-1252817" [ 807.845709] env[61936]: _type = "Task" [ 807.845709] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.852945] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.954895] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 807.975076] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.975726] env[61936]: ERROR nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] Traceback (most recent call last): [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self.driver.spawn(context, instance, image_meta, [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] vm_ref = self.build_virtual_machine(instance, [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.975726] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] for vif in network_info: [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return self._sync_wrapper(fn, *args, **kwargs) [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self.wait() [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self[:] = self._gt.wait() [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return self._exit_event.wait() [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] result = hub.switch() [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 807.976165] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return self.greenlet.switch() [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] result = function(*args, **kwargs) [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] return func(*args, **kwargs) [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] raise e [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] nwinfo = self.network_api.allocate_for_instance( [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] created_port_ids = self._update_ports_for_instance( [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] with excutils.save_and_reraise_exception(): [ 807.976588] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] self.force_reraise() [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] raise self.value [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] updated_port = self._update_port( [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] _ensure_no_port_binding_failure(port) [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] raise exception.PortBindingFailed(port_id=port['id']) [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] nova.exception.PortBindingFailed: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. [ 807.977053] env[61936]: ERROR nova.compute.manager [instance: 4755f16d-254b-47e9-8b50-05df95690445] [ 807.977441] env[61936]: DEBUG nova.compute.utils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.977718] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.924s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.979279] env[61936]: INFO nova.compute.claims [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.981890] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Build of instance 4755f16d-254b-47e9-8b50-05df95690445 was re-scheduled: Binding failed for port df6997b1-d103-47ef-bb00-0393194e7d85, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 807.982359] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 807.982581] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Acquiring lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.983046] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Acquired lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.983046] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.130384] env[61936]: INFO nova.compute.manager [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] [instance: 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85] Took 1.02 seconds to deallocate network for instance. [ 808.179169] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526c53d5-614a-b1b5-fc4d-6120483bcbfc, 'name': SearchDatastore_Task, 'duration_secs': 0.008848} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.179470] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.179696] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.179925] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.180097] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.180287] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.180545] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2025168-aa42-4f42-85a7-559dabad1ccd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.188660] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.188832] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.189558] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbe1fcf2-89a8-48c3-8f47-06f85092a8de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.194686] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 808.194686] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527d8ab2-4e2d-6861-f274-9706ad7aad69" [ 808.194686] env[61936]: _type = "Task" [ 808.194686] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.202471] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527d8ab2-4e2d-6861-f274-9706ad7aad69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.240034] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "1251e7f5-684c-428f-9d63-60b77084d09e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.240314] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "1251e7f5-684c-428f-9d63-60b77084d09e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.240522] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "1251e7f5-684c-428f-9d63-60b77084d09e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.240699] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "1251e7f5-684c-428f-9d63-60b77084d09e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.240862] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "1251e7f5-684c-428f-9d63-60b77084d09e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.243173] env[61936]: INFO nova.compute.manager [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Terminating instance [ 808.260279] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Releasing lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.260625] env[61936]: DEBUG nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Instance network_info: |[{"id": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "address": "fa:16:3e:c2:65:76", "network": {"id": "6a8ce523-f84f-45dc-b062-e340c36e0dc2", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1341537324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b8b985036b44abd9e833f7b99c0e27f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc749785-35", "ovs_interfaceid": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 808.261069] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:65:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4020f51-6e46-4b73-a79e-9fe3fd51b917', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fc749785-35a5-468f-b3a3-b28dd3392c1c', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.269075] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Creating folder: Project (2b8b985036b44abd9e833f7b99c0e27f). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.269658] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4702c11a-344b-4e13-9e60-20082819ec81 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.280058] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Created folder: Project (2b8b985036b44abd9e833f7b99c0e27f) in parent group-v269874. [ 808.280274] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Creating folder: Instances. Parent ref: group-v269907. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.280515] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-875c3d9e-7199-4574-ac27-867a8bda2043 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.289718] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Created folder: Instances in parent group-v269907. [ 808.289964] env[61936]: DEBUG oslo.service.loopingcall [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.290169] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.290400] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d995d38e-d865-4242-a57a-b2856e201515 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.310509] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.310509] env[61936]: value = "task-1252820" [ 808.310509] env[61936]: _type = "Task" [ 808.310509] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.318378] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252820, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.354866] env[61936]: DEBUG oslo_vmware.api [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252817, 'name': PowerOnVM_Task, 'duration_secs': 0.461168} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.355158] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.355377] env[61936]: INFO nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Took 7.34 seconds to spawn the instance on the hypervisor. [ 808.355512] env[61936]: DEBUG nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 808.356854] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1ec9c6-b595-46e2-974b-32143b87c0e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.479472] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.510742] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.549387] env[61936]: DEBUG nova.compute.manager [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Received event network-changed-fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 808.549567] env[61936]: DEBUG nova.compute.manager [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Refreshing instance network info cache due to event network-changed-fc749785-35a5-468f-b3a3-b28dd3392c1c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 808.549731] env[61936]: DEBUG oslo_concurrency.lockutils [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] Acquiring lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.549846] env[61936]: DEBUG oslo_concurrency.lockutils [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] Acquired lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.550066] env[61936]: DEBUG nova.network.neutron [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Refreshing network info cache for port fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 808.606825] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.704372] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527d8ab2-4e2d-6861-f274-9706ad7aad69, 'name': SearchDatastore_Task, 'duration_secs': 0.01075} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.705533] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d5fe9f4-c8e3-43fc-ab03-974d009ab5e7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.710819] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 808.710819] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e57570-3942-29b3-6f22-b2759c37d247" [ 808.710819] env[61936]: _type = "Task" [ 808.710819] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.717895] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e57570-3942-29b3-6f22-b2759c37d247, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.749491] env[61936]: DEBUG nova.compute.manager [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 808.749741] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 808.750586] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585b79c7-377f-4991-9a97-fa0de0dd8030 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.757695] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 808.757986] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d33bce96-525b-48ee-8e91-c960ea1d74a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.763377] env[61936]: DEBUG oslo_vmware.api [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 808.763377] env[61936]: value = "task-1252821" [ 808.763377] env[61936]: _type = "Task" [ 808.763377] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.772615] env[61936]: DEBUG oslo_vmware.api [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.819921] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252820, 'name': CreateVM_Task, 'duration_secs': 0.349373} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.820115] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.820832] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.820996] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.821353] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.821596] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25a7b5b2-e7fb-4239-a1f9-d6dc2f6e06cd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.826054] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 808.826054] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52093757-480d-6f88-6ce4-4ce80aa73612" [ 808.826054] env[61936]: _type = "Task" [ 808.826054] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.834581] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52093757-480d-6f88-6ce4-4ce80aa73612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.877628] env[61936]: INFO nova.compute.manager [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Took 29.38 seconds to build instance. [ 809.110804] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Releasing lock "refresh_cache-4755f16d-254b-47e9-8b50-05df95690445" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.110887] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 809.111093] env[61936]: DEBUG nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 809.111295] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.134799] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.166705] env[61936]: INFO nova.scheduler.client.report [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Deleted allocations for instance 7c9ec7ea-87cd-417a-9695-d43d5cb1eb85 [ 809.230114] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e57570-3942-29b3-6f22-b2759c37d247, 'name': SearchDatastore_Task, 'duration_secs': 0.008728} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.230114] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.230114] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 867a4bdc-8326-48b3-ba17-ca8336533c2d/867a4bdc-8326-48b3-ba17-ca8336533c2d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 809.234158] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3216063d-6d69-43af-868d-ec1492678149 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.247159] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 809.247159] env[61936]: value = "task-1252822" [ 809.247159] env[61936]: _type = "Task" [ 809.247159] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.258918] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.277744] env[61936]: DEBUG oslo_vmware.api [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252821, 'name': PowerOffVM_Task, 'duration_secs': 0.246677} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.280800] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.280800] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.282133] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0b47cc9-cf0d-4bee-a1a3-fd4ab46973ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.321951] env[61936]: DEBUG nova.network.neutron [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Updated VIF entry in instance network info cache for port fc749785-35a5-468f-b3a3-b28dd3392c1c. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 809.322336] env[61936]: DEBUG nova.network.neutron [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Updating instance_info_cache with network_info: [{"id": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "address": "fa:16:3e:c2:65:76", "network": {"id": "6a8ce523-f84f-45dc-b062-e340c36e0dc2", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1341537324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b8b985036b44abd9e833f7b99c0e27f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc749785-35", "ovs_interfaceid": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.342170] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52093757-480d-6f88-6ce4-4ce80aa73612, 'name': SearchDatastore_Task, 'duration_secs': 0.010819} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.342170] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.342170] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.342170] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.342430] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.342430] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.343520] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82746125-21d3-4fe2-8471-43ff3f382413 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.345484] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 809.345664] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 809.345843] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Deleting the datastore file [datastore1] 1251e7f5-684c-428f-9d63-60b77084d09e {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.348964] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97c103a3-8bcb-4b8c-806f-a8c42f5e31d9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.355135] env[61936]: DEBUG oslo_vmware.api [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for the task: (returnval){ [ 809.355135] env[61936]: value = "task-1252824" [ 809.355135] env[61936]: _type = "Task" [ 809.355135] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.366402] env[61936]: DEBUG oslo_vmware.api [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.367477] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.367636] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.368426] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f2a5d36-fd9a-4014-940d-815fbfe174a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.373015] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 809.373015] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e6e160-15f0-06db-8a70-13074c8ec1d2" [ 809.373015] env[61936]: _type = "Task" [ 809.373015] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.380720] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2fda4d77-5b76-4047-a219-e6c6fe51bf97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "144ddda8-db07-4308-8fd3-4045067b5b37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.661s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.385325] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e6e160-15f0-06db-8a70-13074c8ec1d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.407127] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8eaa8a-f992-45c6-8da7-dc049f02a72b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.414360] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f851250-3a8f-495e-bc7e-e66e95972a54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.449079] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db5aa4e-3afb-4169-a7a3-12d559b22218 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.459696] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d14a73d-c25f-422d-b0b2-bc69726131f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.475117] env[61936]: DEBUG nova.compute.provider_tree [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.634704] env[61936]: DEBUG nova.network.neutron [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.677050] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd8ef21f-2709-4749-a45e-4fbece796f9e tempest-VolumesAdminNegativeTest-1273667039 tempest-VolumesAdminNegativeTest-1273667039-project-member] Lock "7c9ec7ea-87cd-417a-9695-d43d5cb1eb85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.854s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.700811] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "144ddda8-db07-4308-8fd3-4045067b5b37" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.701521] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "144ddda8-db07-4308-8fd3-4045067b5b37" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.701815] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "144ddda8-db07-4308-8fd3-4045067b5b37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.702084] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "144ddda8-db07-4308-8fd3-4045067b5b37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.702335] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "144ddda8-db07-4308-8fd3-4045067b5b37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.708987] env[61936]: INFO nova.compute.manager [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Terminating instance [ 809.758440] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252822, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.824962] env[61936]: DEBUG oslo_concurrency.lockutils [req-8e127920-de97-49e7-b626-760a211ef170 req-f5385c00-375d-4abc-bb33-ec824a1afcf4 service nova] Releasing lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.864481] env[61936]: DEBUG oslo_vmware.api [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Task: {'id': task-1252824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.318351} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.864736] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.864924] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 809.865120] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 809.865293] env[61936]: INFO nova.compute.manager [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 809.865523] env[61936]: DEBUG oslo.service.loopingcall [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.865749] env[61936]: DEBUG nova.compute.manager [-] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 809.865850] env[61936]: DEBUG nova.network.neutron [-] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.882666] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e6e160-15f0-06db-8a70-13074c8ec1d2, 'name': SearchDatastore_Task, 'duration_secs': 0.010269} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.883451] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-117247dd-9f71-409c-b756-bd417acd8d15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.886233] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 809.890181] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 809.890181] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521c2467-f807-1282-ace3-2844ac757b13" [ 809.890181] env[61936]: _type = "Task" [ 809.890181] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.899864] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521c2467-f807-1282-ace3-2844ac757b13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.977774] env[61936]: DEBUG nova.scheduler.client.report [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 810.294218] env[61936]: INFO nova.compute.manager [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] [instance: 4755f16d-254b-47e9-8b50-05df95690445] Took 1.03 seconds to deallocate network for instance. [ 810.294218] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 810.294218] env[61936]: DEBUG nova.compute.manager [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 810.294218] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.294218] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133c52ce-8bbe-41dd-a47e-2ebd66b52b15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.294736] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.294736] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b79460ba-7245-4bbe-a81b-c2d4bcc493a0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.294736] env[61936]: DEBUG oslo_vmware.api [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 810.294736] env[61936]: value = "task-1252825" [ 810.294736] env[61936]: _type = "Task" [ 810.294736] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.294736] env[61936]: DEBUG oslo_vmware.api [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.294736] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648627} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.296101] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 867a4bdc-8326-48b3-ba17-ca8336533c2d/867a4bdc-8326-48b3-ba17-ca8336533c2d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 810.296101] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.296101] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de2ae0ac-6d43-431f-90d4-399e8fff1016 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.296101] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 810.296101] env[61936]: value = "task-1252826" [ 810.296101] env[61936]: _type = "Task" [ 810.296101] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.296101] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252826, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.405321] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521c2467-f807-1282-ace3-2844ac757b13, 'name': SearchDatastore_Task, 'duration_secs': 0.015255} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.405640] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.406233] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 301b0504-5cf9-44e0-bd3e-342f8a89278b/301b0504-5cf9-44e0-bd3e-342f8a89278b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.406361] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df61acc2-a873-4728-a454-d7e73c3d8b46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.412437] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 810.412437] env[61936]: value = "task-1252827" [ 810.412437] env[61936]: _type = "Task" [ 810.412437] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.416458] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.421687] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252827, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.482854] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.483419] env[61936]: DEBUG nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 810.486573] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.781s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.486793] env[61936]: DEBUG nova.objects.instance [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 810.699070] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.740220] env[61936]: DEBUG oslo_vmware.api [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252825, 'name': PowerOffVM_Task, 'duration_secs': 0.181692} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.740452] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.740620] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.740878] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54faf6be-a538-4347-b9ef-d83ee2c44f64 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.776043] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252826, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069267} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.776043] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.776381] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f23e76-10df-435d-8128-6272ad2f016d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.799074] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 867a4bdc-8326-48b3-ba17-ca8336533c2d/867a4bdc-8326-48b3-ba17-ca8336533c2d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.799375] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-736a0184-3554-4d13-a4de-def1439a1e00 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.821536] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 810.821536] env[61936]: value = "task-1252829" [ 810.821536] env[61936]: _type = "Task" [ 810.821536] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.826911] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.827238] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.827598] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleting the datastore file [datastore2] 144ddda8-db07-4308-8fd3-4045067b5b37 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.828323] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c23bb04-4358-4d53-9f26-79d75a2deda3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.630338] env[61936]: DEBUG nova.network.neutron [-] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.630338] env[61936]: DEBUG nova.compute.utils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.630338] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252829, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.630338] env[61936]: WARNING oslo_vmware.common.loopingcall [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] task run outlasted interval by 0.264529 sec [ 811.630338] env[61936]: DEBUG nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 811.630338] env[61936]: DEBUG nova.network.neutron [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 811.631269] env[61936]: DEBUG nova.compute.manager [req-927f3578-d8bb-434e-bb1f-fcf0023232bb req-67ee40b8-4ff1-4223-baf6-a07de083156b service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Received event network-vif-deleted-87c99a35-e6ce-45e2-8354-80f5922cea14 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 811.631269] env[61936]: INFO nova.compute.manager [req-927f3578-d8bb-434e-bb1f-fcf0023232bb req-67ee40b8-4ff1-4223-baf6-a07de083156b service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Neutron deleted interface 87c99a35-e6ce-45e2-8354-80f5922cea14; detaching it from the instance and deleting it from the info cache [ 811.631269] env[61936]: DEBUG nova.network.neutron [req-927f3578-d8bb-434e-bb1f-fcf0023232bb req-67ee40b8-4ff1-4223-baf6-a07de083156b service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.631269] env[61936]: DEBUG oslo_vmware.api [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 811.631269] env[61936]: value = "task-1252830" [ 811.631269] env[61936]: _type = "Task" [ 811.631269] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.640166] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252827, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461526} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.643192] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252829, 'name': ReconfigVM_Task, 'duration_secs': 0.31544} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.645872] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 301b0504-5cf9-44e0-bd3e-342f8a89278b/301b0504-5cf9-44e0-bd3e-342f8a89278b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.646447] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.647694] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 867a4bdc-8326-48b3-ba17-ca8336533c2d/867a4bdc-8326-48b3-ba17-ca8336533c2d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.651908] env[61936]: INFO nova.scheduler.client.report [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Deleted allocations for instance 4755f16d-254b-47e9-8b50-05df95690445 [ 811.666052] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ca79b0f-bce7-41c1-9087-46aa7a0ccbb5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.673280] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00dd1fd9-a83e-4458-bb99-fcd2adcf8280 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.687062] env[61936]: DEBUG nova.policy [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9260b79bb7f4e44bb1f50b16273df09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ce9d789b30c49758bd073d17b3f1281', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 811.701705] env[61936]: DEBUG oslo_vmware.api [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.712414] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 811.712414] env[61936]: value = "task-1252831" [ 811.712414] env[61936]: _type = "Task" [ 811.712414] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.712414] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 811.712414] env[61936]: value = "task-1252832" [ 811.712414] env[61936]: _type = "Task" [ 811.712414] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.731356] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252831, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.731700] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.049876] env[61936]: DEBUG nova.network.neutron [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Successfully created port: a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.088394] env[61936]: INFO nova.compute.manager [-] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Took 2.22 seconds to deallocate network for instance. [ 812.091990] env[61936]: DEBUG nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 812.096368] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0f1173f-7d51-4ad9-afbe-ed1a321568cb tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.610s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.099161] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.117s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.112281] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7e574f98-0e41-4d15-9e2f-08075a1ae7d9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.125964] env[61936]: DEBUG oslo_vmware.api [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1252830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140388} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.127756] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 812.127958] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 812.131112] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 812.131112] env[61936]: INFO nova.compute.manager [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Took 1.91 seconds to destroy the instance on the hypervisor. [ 812.131112] env[61936]: DEBUG oslo.service.loopingcall [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.131112] env[61936]: DEBUG nova.compute.manager [-] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 812.131112] env[61936]: DEBUG nova.network.neutron [-] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 812.135635] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547a2745-f4af-44c9-b18c-daaf28549011 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.163895] env[61936]: DEBUG nova.compute.manager [req-927f3578-d8bb-434e-bb1f-fcf0023232bb req-67ee40b8-4ff1-4223-baf6-a07de083156b service nova] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Detach interface failed, port_id=87c99a35-e6ce-45e2-8354-80f5922cea14, reason: Instance 1251e7f5-684c-428f-9d63-60b77084d09e could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 812.192685] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d9ed7b71-d116-4501-99c8-57cced62bbbd tempest-InstanceActionsTestJSON-197586485 tempest-InstanceActionsTestJSON-197586485-project-member] Lock "4755f16d-254b-47e9-8b50-05df95690445" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.555s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.219620] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252831, 'name': Rename_Task, 'duration_secs': 0.141296} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.220976] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.221630] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071227} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.225033] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff117094-833f-4aaa-8be5-e88f7ccdd831 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.225737] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.226608] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c81f76-33ca-4734-b24e-b12ecbb30ff8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.251198] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 301b0504-5cf9-44e0-bd3e-342f8a89278b/301b0504-5cf9-44e0-bd3e-342f8a89278b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.255017] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67d4d157-a356-4b47-8d68-f280ee5e35e7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.268300] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 812.268300] env[61936]: value = "task-1252833" [ 812.268300] env[61936]: _type = "Task" [ 812.268300] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.274140] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 812.274140] env[61936]: value = "task-1252834" [ 812.274140] env[61936]: _type = "Task" [ 812.274140] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.279294] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252833, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.287703] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252834, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.621601] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.696722] env[61936]: DEBUG nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 812.783555] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252833, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.792708] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252834, 'name': ReconfigVM_Task, 'duration_secs': 0.268365} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.795437] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 301b0504-5cf9-44e0-bd3e-342f8a89278b/301b0504-5cf9-44e0-bd3e-342f8a89278b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.796308] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-859050b1-289f-4874-9192-deb5b545f85e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.802783] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 812.802783] env[61936]: value = "task-1252835" [ 812.802783] env[61936]: _type = "Task" [ 812.802783] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.813194] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252835, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.842727] env[61936]: DEBUG nova.compute.manager [req-50a5bd82-eaad-4d49-a9e9-0c65e8a46bad req-6b8148af-3bb5-4915-b87e-05553e07eff5 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Received event network-vif-deleted-670384d2-16a5-4c63-a597-a1513479b68b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 812.842962] env[61936]: INFO nova.compute.manager [req-50a5bd82-eaad-4d49-a9e9-0c65e8a46bad req-6b8148af-3bb5-4915-b87e-05553e07eff5 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Neutron deleted interface 670384d2-16a5-4c63-a597-a1513479b68b; detaching it from the instance and deleting it from the info cache [ 812.843231] env[61936]: DEBUG nova.network.neutron [req-50a5bd82-eaad-4d49-a9e9-0c65e8a46bad req-6b8148af-3bb5-4915-b87e-05553e07eff5 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.991598] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0a01c9-de60-4f63-afef-02a12ed0787e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.999940] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ea8b4c-7d05-4d70-b73b-c8701f80646f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.032801] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587e0748-f795-4674-8915-7765ca20d126 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.040952] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2605d7b2-8076-43ec-b8cf-d37b0f8b3985 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.054693] env[61936]: DEBUG nova.compute.provider_tree [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.105133] env[61936]: DEBUG nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 813.122646] env[61936]: DEBUG nova.network.neutron [-] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.140795] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 813.141047] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.141206] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 813.141384] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.141526] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 813.141669] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 813.141865] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 813.142128] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 813.142368] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 813.142594] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 813.142880] env[61936]: DEBUG nova.virt.hardware [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 813.144060] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ff9fee-d688-4a68-bc4b-1cf935e1abce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.152670] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f955b29e-7588-4f85-a316-4efd48f609d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.220680] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.279872] env[61936]: DEBUG oslo_vmware.api [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252833, 'name': PowerOnVM_Task, 'duration_secs': 0.548009} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.280216] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.280484] env[61936]: INFO nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Took 9.78 seconds to spawn the instance on the hypervisor. [ 813.280671] env[61936]: DEBUG nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 813.281450] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd62952-8d09-46be-95ce-2cadfeb7c651 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.312318] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252835, 'name': Rename_Task, 'duration_secs': 0.143949} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.312623] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.312887] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dfc0998c-c133-4065-a1a3-cc4e0d974817 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.321063] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 813.321063] env[61936]: value = "task-1252836" [ 813.321063] env[61936]: _type = "Task" [ 813.321063] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.331803] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252836, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.348413] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36bc8652-1f69-47b8-aec0-c9dcca435136 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.357483] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a964a341-4b08-49fa-9f54-d82298bfa436 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.383393] env[61936]: DEBUG nova.compute.manager [req-50a5bd82-eaad-4d49-a9e9-0c65e8a46bad req-6b8148af-3bb5-4915-b87e-05553e07eff5 service nova] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Detach interface failed, port_id=670384d2-16a5-4c63-a597-a1513479b68b, reason: Instance 144ddda8-db07-4308-8fd3-4045067b5b37 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 813.482205] env[61936]: DEBUG nova.compute.manager [req-4211b190-6283-409f-9cdb-2cc05596ccc4 req-29e1dfed-60f1-41b6-bf5b-724ae5ce1a20 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Received event network-vif-plugged-a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 813.482433] env[61936]: DEBUG oslo_concurrency.lockutils [req-4211b190-6283-409f-9cdb-2cc05596ccc4 req-29e1dfed-60f1-41b6-bf5b-724ae5ce1a20 service nova] Acquiring lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.482599] env[61936]: DEBUG oslo_concurrency.lockutils [req-4211b190-6283-409f-9cdb-2cc05596ccc4 req-29e1dfed-60f1-41b6-bf5b-724ae5ce1a20 service nova] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.482765] env[61936]: DEBUG oslo_concurrency.lockutils [req-4211b190-6283-409f-9cdb-2cc05596ccc4 req-29e1dfed-60f1-41b6-bf5b-724ae5ce1a20 service nova] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.482925] env[61936]: DEBUG nova.compute.manager [req-4211b190-6283-409f-9cdb-2cc05596ccc4 req-29e1dfed-60f1-41b6-bf5b-724ae5ce1a20 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] No waiting events found dispatching network-vif-plugged-a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 813.483565] env[61936]: WARNING nova.compute.manager [req-4211b190-6283-409f-9cdb-2cc05596ccc4 req-29e1dfed-60f1-41b6-bf5b-724ae5ce1a20 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Received unexpected event network-vif-plugged-a342f5d6-5a18-4e6f-b046-fd7365684711 for instance with vm_state building and task_state spawning. [ 813.557551] env[61936]: DEBUG nova.scheduler.client.report [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 813.579624] env[61936]: DEBUG nova.network.neutron [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Successfully updated port: a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 813.628709] env[61936]: INFO nova.compute.manager [-] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Took 1.50 seconds to deallocate network for instance. [ 813.802394] env[61936]: INFO nova.compute.manager [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Took 29.76 seconds to build instance. [ 813.836030] env[61936]: DEBUG oslo_vmware.api [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252836, 'name': PowerOnVM_Task, 'duration_secs': 0.461316} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.836306] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 813.836496] env[61936]: INFO nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Took 7.87 seconds to spawn the instance on the hypervisor. [ 813.836665] env[61936]: DEBUG nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 813.837456] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c48f5f8-6a60-4712-b850-648b4454787d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.063721] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.964s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.064198] env[61936]: ERROR nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Traceback (most recent call last): [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self.driver.spawn(context, instance, image_meta, [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] vm_ref = self.build_virtual_machine(instance, [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.064198] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] for vif in network_info: [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return self._sync_wrapper(fn, *args, **kwargs) [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self.wait() [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self[:] = self._gt.wait() [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return self._exit_event.wait() [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] result = hub.switch() [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 814.064521] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return self.greenlet.switch() [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] result = function(*args, **kwargs) [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] return func(*args, **kwargs) [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] raise e [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] nwinfo = self.network_api.allocate_for_instance( [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] created_port_ids = self._update_ports_for_instance( [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] with excutils.save_and_reraise_exception(): [ 814.064789] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] self.force_reraise() [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] raise self.value [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] updated_port = self._update_port( [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] _ensure_no_port_binding_failure(port) [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] raise exception.PortBindingFailed(port_id=port['id']) [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] nova.exception.PortBindingFailed: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. [ 814.065053] env[61936]: ERROR nova.compute.manager [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] [ 814.065275] env[61936]: DEBUG nova.compute.utils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.067216] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Build of instance 2fdb605e-72d5-4eef-bab2-0601e730243a was re-scheduled: Binding failed for port 462018e4-3904-4c95-8328-7819e8709198, please check neutron logs for more information. {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 814.067889] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Unplugging VIFs for instance {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 814.067979] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.068184] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquired lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.068412] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.069717] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.055s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.071180] env[61936]: INFO nova.compute.claims [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.083435] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.083606] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.083778] env[61936]: DEBUG nova.network.neutron [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.137601] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.307786] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1f0ed2ed-5708-41a8-bf7c-65da89459c31 tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.752s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.371650] env[61936]: INFO nova.compute.manager [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Took 28.37 seconds to build instance. [ 814.497820] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "867a4bdc-8326-48b3-ba17-ca8336533c2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.497902] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.498091] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "867a4bdc-8326-48b3-ba17-ca8336533c2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.498272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.498450] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.503028] env[61936]: INFO nova.compute.manager [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Terminating instance [ 814.611162] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.642093] env[61936]: DEBUG nova.network.neutron [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.698165] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.810298] env[61936]: DEBUG nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 814.873805] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad35800a-9169-4347-beee-1ae9038df4cf tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.094s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.964077] env[61936]: DEBUG nova.network.neutron [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updating instance_info_cache with network_info: [{"id": "a342f5d6-5a18-4e6f-b046-fd7365684711", "address": "fa:16:3e:93:27:c5", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa342f5d6-5a", "ovs_interfaceid": "a342f5d6-5a18-4e6f-b046-fd7365684711", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.006608] env[61936]: DEBUG nova.compute.manager [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 815.006844] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.007757] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51293683-2357-4327-9499-d29c87e51598 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.016812] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.017081] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a51fd516-957b-4efa-a031-40ce624fef21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.023350] env[61936]: DEBUG oslo_vmware.api [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 815.023350] env[61936]: value = "task-1252837" [ 815.023350] env[61936]: _type = "Task" [ 815.023350] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.041928] env[61936]: DEBUG oslo_vmware.api [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252837, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.201687] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Releasing lock "refresh_cache-2fdb605e-72d5-4eef-bab2-0601e730243a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.202021] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61936) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 815.202232] env[61936]: DEBUG nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 815.202403] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.244743] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.327715] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.376340] env[61936]: DEBUG nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 815.469397] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.470285] env[61936]: DEBUG nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Instance network_info: |[{"id": "a342f5d6-5a18-4e6f-b046-fd7365684711", "address": "fa:16:3e:93:27:c5", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa342f5d6-5a", "ovs_interfaceid": "a342f5d6-5a18-4e6f-b046-fd7365684711", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 815.470733] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:27:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a342f5d6-5a18-4e6f-b046-fd7365684711', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.480871] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Creating folder: Project (9ce9d789b30c49758bd073d17b3f1281). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.481887] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a3fe915-cfee-4c34-ba61-3d7238a3d13c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.486497] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fda13f-9319-4968-abc2-436b69ddf478 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.494626] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74c000f-2488-48de-b1e8-ffd14da48ab2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.498888] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Created folder: Project (9ce9d789b30c49758bd073d17b3f1281) in parent group-v269874. [ 815.499144] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Creating folder: Instances. Parent ref: group-v269910. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.499733] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a2e8b22-b8d7-4d7f-8953-48e76de874c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.536510] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaaff05-8e0b-468b-93fe-1eedf5f5760d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.538989] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Created folder: Instances in parent group-v269910. [ 815.539297] env[61936]: DEBUG oslo.service.loopingcall [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.539738] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.540716] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a4fa2bc-08a9-41c2-9571-9ae25d4ee76f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.561231] env[61936]: DEBUG oslo_vmware.api [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252837, 'name': PowerOffVM_Task, 'duration_secs': 0.204742} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.562955] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960ea817-a3c3-46c1-8072-2a4cb0d0b609 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.566863] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.567118] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.568505] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8cff1eb-3878-4372-bebb-8f8383b688b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.569829] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.569829] env[61936]: value = "task-1252840" [ 815.569829] env[61936]: _type = "Task" [ 815.569829] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.582550] env[61936]: DEBUG nova.compute.provider_tree [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.588057] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252840, 'name': CreateVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.601669] env[61936]: DEBUG nova.compute.manager [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Received event network-changed-a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 815.601886] env[61936]: DEBUG nova.compute.manager [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Refreshing instance network info cache due to event network-changed-a342f5d6-5a18-4e6f-b046-fd7365684711. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 815.602107] env[61936]: DEBUG oslo_concurrency.lockutils [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] Acquiring lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.602250] env[61936]: DEBUG oslo_concurrency.lockutils [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] Acquired lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.602402] env[61936]: DEBUG nova.network.neutron [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Refreshing network info cache for port a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 815.641189] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.642629] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.642629] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Deleting the datastore file [datastore2] 867a4bdc-8326-48b3-ba17-ca8336533c2d {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.642629] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-abd0b3d1-677d-4180-a2ba-348fc933f17f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.649364] env[61936]: DEBUG oslo_vmware.api [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for the task: (returnval){ [ 815.649364] env[61936]: value = "task-1252842" [ 815.649364] env[61936]: _type = "Task" [ 815.649364] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.658189] env[61936]: DEBUG oslo_vmware.api [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.747299] env[61936]: DEBUG nova.network.neutron [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.904903] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.081750] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252840, 'name': CreateVM_Task, 'duration_secs': 0.327587} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.082426] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.084279] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.084630] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.085018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.086025] env[61936]: DEBUG nova.scheduler.client.report [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 816.089850] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78efc81c-9932-446f-9e78-cc3ca2ff6587 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.095350] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 816.095350] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5294948a-79cb-8816-8a08-d1993e197624" [ 816.095350] env[61936]: _type = "Task" [ 816.095350] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.104014] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5294948a-79cb-8816-8a08-d1993e197624, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.163218] env[61936]: DEBUG oslo_vmware.api [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Task: {'id': task-1252842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127318} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.163656] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.163951] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.164265] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.164553] env[61936]: INFO nova.compute.manager [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 816.165486] env[61936]: DEBUG oslo.service.loopingcall [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.165486] env[61936]: DEBUG nova.compute.manager [-] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 816.165486] env[61936]: DEBUG nova.network.neutron [-] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.250180] env[61936]: INFO nova.compute.manager [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 2fdb605e-72d5-4eef-bab2-0601e730243a] Took 1.05 seconds to deallocate network for instance. [ 816.396602] env[61936]: DEBUG nova.network.neutron [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updated VIF entry in instance network info cache for port a342f5d6-5a18-4e6f-b046-fd7365684711. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 816.396602] env[61936]: DEBUG nova.network.neutron [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updating instance_info_cache with network_info: [{"id": "a342f5d6-5a18-4e6f-b046-fd7365684711", "address": "fa:16:3e:93:27:c5", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa342f5d6-5a", "ovs_interfaceid": "a342f5d6-5a18-4e6f-b046-fd7365684711", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.593150] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.593837] env[61936]: DEBUG nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 816.599196] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.492s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.599196] env[61936]: DEBUG nova.objects.instance [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lazy-loading 'resources' on Instance uuid 498a77d6-ca0c-4841-b4a2-b7a024281c6a {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.610153] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5294948a-79cb-8816-8a08-d1993e197624, 'name': SearchDatastore_Task, 'duration_secs': 0.009866} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.611391] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.611391] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.611391] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.611647] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.611723] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.612105] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-276fc6a0-07d9-421d-a012-948700e00146 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.621552] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.621767] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.622539] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6366b7ed-7b0d-46f3-a778-7e169f909c23 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.628262] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 816.628262] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b0c99c-dcb2-3f2b-70e7-07c17bdd9843" [ 816.628262] env[61936]: _type = "Task" [ 816.628262] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.644772] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b0c99c-dcb2-3f2b-70e7-07c17bdd9843, 'name': SearchDatastore_Task, 'duration_secs': 0.009801} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.645630] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e3c1b64-4791-40b3-918b-b9e1748e5160 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.653926] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 816.653926] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d88de5-eae2-922a-4099-b88bcd822da6" [ 816.653926] env[61936]: _type = "Task" [ 816.653926] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.662574] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d88de5-eae2-922a-4099-b88bcd822da6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.903408] env[61936]: DEBUG oslo_concurrency.lockutils [req-fdb52348-3ae3-45a1-9684-233ec71e6383 req-ac22008e-1b7a-41d7-bea8-c0ed58de1b0a service nova] Releasing lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.105245] env[61936]: DEBUG nova.compute.utils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.110272] env[61936]: DEBUG nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 817.110272] env[61936]: DEBUG nova.network.neutron [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.138210] env[61936]: DEBUG nova.network.neutron [-] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.168585] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d88de5-eae2-922a-4099-b88bcd822da6, 'name': SearchDatastore_Task, 'duration_secs': 0.008671} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.168585] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.168757] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377/7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.169397] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d2a89a4-f2d5-4be0-aa89-3ca8147ae10f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.176875] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 817.176875] env[61936]: value = "task-1252843" [ 817.176875] env[61936]: _type = "Task" [ 817.176875] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.188101] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252843, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.289280] env[61936]: INFO nova.scheduler.client.report [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Deleted allocations for instance 2fdb605e-72d5-4eef-bab2-0601e730243a [ 817.326386] env[61936]: DEBUG nova.policy [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b05bdcace008487dba580405eb672ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90c29449db6c4beaa7f190d9225cb08a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.519748] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7070bdea-a39a-4b9e-b0bd-e8ac25fe7089 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.529757] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e408e86-a4f1-4f0f-bc07-46765b24f3cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.570064] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e100d469-b499-4540-bca0-b40a0fbc71c5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.583487] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b44d84-d4eb-410a-a970-bc1b88720c0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.602342] env[61936]: DEBUG nova.compute.provider_tree [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.610660] env[61936]: DEBUG nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 817.641069] env[61936]: INFO nova.compute.manager [-] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Took 1.48 seconds to deallocate network for instance. [ 817.689676] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252843, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445262} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.689676] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377/7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 817.689676] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.689676] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e161ba3-be0d-494c-91fa-ff1463b89f8b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.695718] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 817.695718] env[61936]: value = "task-1252844" [ 817.695718] env[61936]: _type = "Task" [ 817.695718] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.703686] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.788817] env[61936]: DEBUG nova.compute.manager [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Received event network-changed-fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 817.788957] env[61936]: DEBUG nova.compute.manager [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Refreshing instance network info cache due to event network-changed-fc749785-35a5-468f-b3a3-b28dd3392c1c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 817.789233] env[61936]: DEBUG oslo_concurrency.lockutils [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] Acquiring lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.789412] env[61936]: DEBUG oslo_concurrency.lockutils [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] Acquired lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.790621] env[61936]: DEBUG nova.network.neutron [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Refreshing network info cache for port fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.808992] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb6153c8-1e77-4b8e-b583-b86bf7c7ce6d tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "2fdb605e-72d5-4eef-bab2-0601e730243a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.007s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.814718] env[61936]: DEBUG nova.network.neutron [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Successfully created port: 7553572a-bf60-44ff-a283-6ab783e30788 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.106809] env[61936]: DEBUG nova.scheduler.client.report [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 818.149620] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.206707] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064667} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.207990] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.208818] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eae7e8a-adbf-477a-aaf9-d59be7b06a2b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.236210] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377/7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.236377] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b76833b-4854-4645-9ec2-46c1dac77ba6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.259367] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 818.259367] env[61936]: value = "task-1252845" [ 818.259367] env[61936]: _type = "Task" [ 818.259367] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.269744] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252845, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.311906] env[61936]: DEBUG nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 818.356818] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.357089] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.357781] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.357781] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.357781] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.360257] env[61936]: INFO nova.compute.manager [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Terminating instance [ 818.613955] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.016s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.616246] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.074s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.624392] env[61936]: DEBUG nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 818.642016] env[61936]: INFO nova.scheduler.client.report [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Deleted allocations for instance 498a77d6-ca0c-4841-b4a2-b7a024281c6a [ 818.667094] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 818.667094] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.667094] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 818.667094] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.667253] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 818.667311] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 818.667529] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 818.667659] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 818.667819] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 818.672287] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 818.672287] env[61936]: DEBUG nova.virt.hardware [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 818.673177] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877669ee-3c60-4cb3-ae79-d80072da2509 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.684148] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a908ef1-9366-4211-91da-af7e00a37e25 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.775537] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252845, 'name': ReconfigVM_Task, 'duration_secs': 0.327345} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.775537] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377/7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.775962] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3dc7a51-9194-4962-9a93-a6582c81dc88 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.782733] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 818.782733] env[61936]: value = "task-1252846" [ 818.782733] env[61936]: _type = "Task" [ 818.782733] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.790725] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252846, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.843196] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.867176] env[61936]: DEBUG nova.compute.manager [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 818.867423] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.869459] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e56afc5-57a8-4757-b79f-8aadf256ac78 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.878303] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.878303] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8187c0c-0cb7-48e6-a123-e95ea67b2eaa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.882711] env[61936]: DEBUG oslo_vmware.api [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 818.882711] env[61936]: value = "task-1252847" [ 818.882711] env[61936]: _type = "Task" [ 818.882711] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.891729] env[61936]: DEBUG oslo_vmware.api [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.937705] env[61936]: DEBUG nova.network.neutron [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Updated VIF entry in instance network info cache for port fc749785-35a5-468f-b3a3-b28dd3392c1c. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 818.938231] env[61936]: DEBUG nova.network.neutron [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Updating instance_info_cache with network_info: [{"id": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "address": "fa:16:3e:c2:65:76", "network": {"id": "6a8ce523-f84f-45dc-b062-e340c36e0dc2", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1341537324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.173", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b8b985036b44abd9e833f7b99c0e27f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc749785-35", "ovs_interfaceid": "fc749785-35a5-468f-b3a3-b28dd3392c1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.148438] env[61936]: DEBUG oslo_concurrency.lockutils [None req-61da53cd-d927-4bd1-9673-15f4204a770c tempest-ServerShowV254Test-545134158 tempest-ServerShowV254Test-545134158-project-member] Lock "498a77d6-ca0c-4841-b4a2-b7a024281c6a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.239s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.293068] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252846, 'name': Rename_Task, 'duration_secs': 0.139697} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.293409] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.293615] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d7ee592-9bdc-4dcb-a893-34a2ef167ea6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.300854] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 819.300854] env[61936]: value = "task-1252848" [ 819.300854] env[61936]: _type = "Task" [ 819.300854] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.310665] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252848, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.394575] env[61936]: DEBUG oslo_vmware.api [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252847, 'name': PowerOffVM_Task, 'duration_secs': 0.191034} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.394575] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.395506] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 819.395506] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c2f6c0d-d1f9-47da-b841-f3fb8236d5e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.441402] env[61936]: DEBUG oslo_concurrency.lockutils [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] Releasing lock "refresh_cache-301b0504-5cf9-44e0-bd3e-342f8a89278b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.441674] env[61936]: DEBUG nova.compute.manager [req-684dae10-09ea-49e8-b9c5-b0bf53945187 req-2fd3ef9a-51a2-422d-a364-6b6524e0ee7f service nova] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Received event network-vif-deleted-c319d6d8-72f1-4705-8ee6-1603b01cf344 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 819.656172] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 50cb3f79-39b2-46ab-a35c-fdad3bb266cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.656172] env[61936]: WARNING nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 1251e7f5-684c-428f-9d63-60b77084d09e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.656172] env[61936]: WARNING nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 144ddda8-db07-4308-8fd3-4045067b5b37 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.656172] env[61936]: WARNING nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 867a4bdc-8326-48b3-ba17-ca8336533c2d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 819.656172] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 301b0504-5cf9-44e0-bd3e-342f8a89278b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.656370] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.656370] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance d410e029-cb32-4cc6-b11d-9cba643e3b7b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 819.806205] env[61936]: DEBUG nova.network.neutron [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Successfully updated port: 7553572a-bf60-44ff-a283-6ab783e30788 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.821799] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252848, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.829962] env[61936]: DEBUG nova.compute.manager [req-c6f40fb2-2fff-48b3-b2f9-4547786f27fa req-a17649db-aedf-4e90-a3f2-b1606d6688b1 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Received event network-vif-plugged-7553572a-bf60-44ff-a283-6ab783e30788 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 819.830305] env[61936]: DEBUG oslo_concurrency.lockutils [req-c6f40fb2-2fff-48b3-b2f9-4547786f27fa req-a17649db-aedf-4e90-a3f2-b1606d6688b1 service nova] Acquiring lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.830560] env[61936]: DEBUG oslo_concurrency.lockutils [req-c6f40fb2-2fff-48b3-b2f9-4547786f27fa req-a17649db-aedf-4e90-a3f2-b1606d6688b1 service nova] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.830792] env[61936]: DEBUG oslo_concurrency.lockutils [req-c6f40fb2-2fff-48b3-b2f9-4547786f27fa req-a17649db-aedf-4e90-a3f2-b1606d6688b1 service nova] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.831046] env[61936]: DEBUG nova.compute.manager [req-c6f40fb2-2fff-48b3-b2f9-4547786f27fa req-a17649db-aedf-4e90-a3f2-b1606d6688b1 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] No waiting events found dispatching network-vif-plugged-7553572a-bf60-44ff-a283-6ab783e30788 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 819.831320] env[61936]: WARNING nova.compute.manager [req-c6f40fb2-2fff-48b3-b2f9-4547786f27fa req-a17649db-aedf-4e90-a3f2-b1606d6688b1 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Received unexpected event network-vif-plugged-7553572a-bf60-44ff-a283-6ab783e30788 for instance with vm_state building and task_state spawning. [ 820.160593] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 5526cbd5-b1ad-453b-8401-eee7aa356606 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 820.172512] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.173365] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.173699] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Deleting the datastore file [datastore1] 50cb3f79-39b2-46ab-a35c-fdad3bb266cd {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.174828] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-769d13be-837b-4ab5-9797-23ebba64ac95 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.185311] env[61936]: DEBUG oslo_vmware.api [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for the task: (returnval){ [ 820.185311] env[61936]: value = "task-1252850" [ 820.185311] env[61936]: _type = "Task" [ 820.185311] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.196339] env[61936]: DEBUG oslo_vmware.api [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.308194] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-d410e029-cb32-4cc6-b11d-9cba643e3b7b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.308596] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-d410e029-cb32-4cc6-b11d-9cba643e3b7b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.308852] env[61936]: DEBUG nova.network.neutron [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.324507] env[61936]: DEBUG oslo_vmware.api [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252848, 'name': PowerOnVM_Task, 'duration_secs': 0.646698} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.325739] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.326096] env[61936]: INFO nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Took 7.22 seconds to spawn the instance on the hypervisor. [ 820.326400] env[61936]: DEBUG nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 820.327756] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8670ffcf-c333-4917-81cc-3bc1a1b71fc4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.663784] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance bdea3742-5789-4e97-b139-33cfee134843 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 820.694191] env[61936]: DEBUG oslo_vmware.api [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Task: {'id': task-1252850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168954} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.694443] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.694618] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.694786] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.694948] env[61936]: INFO nova.compute.manager [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Took 1.83 seconds to destroy the instance on the hypervisor. [ 820.695282] env[61936]: DEBUG oslo.service.loopingcall [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.695848] env[61936]: DEBUG nova.compute.manager [-] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 820.695948] env[61936]: DEBUG nova.network.neutron [-] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.851411] env[61936]: INFO nova.compute.manager [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Took 28.82 seconds to build instance. [ 820.873212] env[61936]: DEBUG nova.network.neutron [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.136493] env[61936]: DEBUG nova.network.neutron [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Updating instance_info_cache with network_info: [{"id": "7553572a-bf60-44ff-a283-6ab783e30788", "address": "fa:16:3e:54:e4:aa", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7553572a-bf", "ovs_interfaceid": "7553572a-bf60-44ff-a283-6ab783e30788", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.167631] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 77a59650-d28b-4862-bb35-6910e6f7d6c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 821.355674] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a575ebce-e616-484c-bd7c-f7734a031545 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.454s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.454863] env[61936]: DEBUG nova.network.neutron [-] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.638357] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-d410e029-cb32-4cc6-b11d-9cba643e3b7b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.638751] env[61936]: DEBUG nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Instance network_info: |[{"id": "7553572a-bf60-44ff-a283-6ab783e30788", "address": "fa:16:3e:54:e4:aa", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7553572a-bf", "ovs_interfaceid": "7553572a-bf60-44ff-a283-6ab783e30788", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 821.639240] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:e4:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7553572a-bf60-44ff-a283-6ab783e30788', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.652969] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating folder: Project (90c29449db6c4beaa7f190d9225cb08a). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.653407] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e42876fb-36f0-42aa-a161-840d88d1ac73 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.666199] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created folder: Project (90c29449db6c4beaa7f190d9225cb08a) in parent group-v269874. [ 821.666487] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating folder: Instances. Parent ref: group-v269913. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.666835] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18b78dce-04f8-4983-a8a7-a3a45ecc78ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.671360] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fafc995e-4c93-4ca0-b078-24a1ae0ab427 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 821.676904] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created folder: Instances in parent group-v269913. [ 821.677158] env[61936]: DEBUG oslo.service.loopingcall [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.677348] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.677551] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd437c60-0e90-4c04-8618-a55d45c564d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.699053] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.699053] env[61936]: value = "task-1252853" [ 821.699053] env[61936]: _type = "Task" [ 821.699053] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.707242] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252853, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.817655] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.817892] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.859030] env[61936]: DEBUG nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 821.872799] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.873040] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.903987] env[61936]: DEBUG nova.compute.manager [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Received event network-changed-7553572a-bf60-44ff-a283-6ab783e30788 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 821.903987] env[61936]: DEBUG nova.compute.manager [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Refreshing instance network info cache due to event network-changed-7553572a-bf60-44ff-a283-6ab783e30788. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 821.904553] env[61936]: DEBUG oslo_concurrency.lockutils [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] Acquiring lock "refresh_cache-d410e029-cb32-4cc6-b11d-9cba643e3b7b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.904752] env[61936]: DEBUG oslo_concurrency.lockutils [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] Acquired lock "refresh_cache-d410e029-cb32-4cc6-b11d-9cba643e3b7b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.904862] env[61936]: DEBUG nova.network.neutron [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Refreshing network info cache for port 7553572a-bf60-44ff-a283-6ab783e30788 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 821.958968] env[61936]: INFO nova.compute.manager [-] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Took 1.26 seconds to deallocate network for instance. [ 822.174435] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 822.211295] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252853, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.390624] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.466609] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.677271] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e5e3e7af-8104-4f7a-ae18-81e03a932608 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 822.716290] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252853, 'name': CreateVM_Task, 'duration_secs': 0.808211} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.720309] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.720309] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.720309] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.720309] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.720309] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6a2bd65-309b-490b-9089-1db87e4e5e50 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.725325] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 822.725325] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5245e308-de89-9a3e-7db2-0fd9ed161ed8" [ 822.725325] env[61936]: _type = "Task" [ 822.725325] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.735752] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5245e308-de89-9a3e-7db2-0fd9ed161ed8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.939247] env[61936]: DEBUG nova.network.neutron [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Updated VIF entry in instance network info cache for port 7553572a-bf60-44ff-a283-6ab783e30788. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.939247] env[61936]: DEBUG nova.network.neutron [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Updating instance_info_cache with network_info: [{"id": "7553572a-bf60-44ff-a283-6ab783e30788", "address": "fa:16:3e:54:e4:aa", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7553572a-bf", "ovs_interfaceid": "7553572a-bf60-44ff-a283-6ab783e30788", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.040322] env[61936]: DEBUG nova.compute.manager [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Received event network-changed-a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 823.040523] env[61936]: DEBUG nova.compute.manager [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Refreshing instance network info cache due to event network-changed-a342f5d6-5a18-4e6f-b046-fd7365684711. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 823.040741] env[61936]: DEBUG oslo_concurrency.lockutils [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] Acquiring lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.040897] env[61936]: DEBUG oslo_concurrency.lockutils [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] Acquired lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.041080] env[61936]: DEBUG nova.network.neutron [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Refreshing network info cache for port a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.181826] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f2844173-4114-4158-99ad-ac651efbb607 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 823.238026] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5245e308-de89-9a3e-7db2-0fd9ed161ed8, 'name': SearchDatastore_Task, 'duration_secs': 0.01111} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.238026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.238026] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.238026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.238317] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.238620] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.238978] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce189043-f5a3-4b3e-938c-e4bef45891e6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.247525] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.247829] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.248667] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff743435-c04b-494a-b045-520999c475ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.254191] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 823.254191] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b9281a-b8f0-c9b7-6962-30ffec304166" [ 823.254191] env[61936]: _type = "Task" [ 823.254191] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.263194] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b9281a-b8f0-c9b7-6962-30ffec304166, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.441719] env[61936]: DEBUG oslo_concurrency.lockutils [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] Releasing lock "refresh_cache-d410e029-cb32-4cc6-b11d-9cba643e3b7b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.442012] env[61936]: DEBUG nova.compute.manager [req-0b52763d-5e0b-4610-bb73-2ea48e52c48b req-27698ff7-1c40-427f-b0b4-93576e746685 service nova] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Received event network-vif-deleted-1056fca0-7eeb-464b-befc-9a19819b7a58 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 823.684343] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e800f5e9-deb0-4592-aded-3768d238e311 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 823.764748] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b9281a-b8f0-c9b7-6962-30ffec304166, 'name': SearchDatastore_Task, 'duration_secs': 0.009251} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.765513] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e323ecc-d82e-4253-b11e-e8f7cb17e379 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.771753] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 823.771753] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520debfa-1827-701f-cca7-a904020e1e36" [ 823.771753] env[61936]: _type = "Task" [ 823.771753] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.778852] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520debfa-1827-701f-cca7-a904020e1e36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.862390] env[61936]: DEBUG nova.network.neutron [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updated VIF entry in instance network info cache for port a342f5d6-5a18-4e6f-b046-fd7365684711. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.862652] env[61936]: DEBUG nova.network.neutron [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updating instance_info_cache with network_info: [{"id": "a342f5d6-5a18-4e6f-b046-fd7365684711", "address": "fa:16:3e:93:27:c5", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa342f5d6-5a", "ovs_interfaceid": "a342f5d6-5a18-4e6f-b046-fd7365684711", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.188574] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e15be950-1c32-4801-b4e4-a4d31ff0cfec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 824.281576] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520debfa-1827-701f-cca7-a904020e1e36, 'name': SearchDatastore_Task, 'duration_secs': 0.043933} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.281773] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.282033] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] d410e029-cb32-4cc6-b11d-9cba643e3b7b/d410e029-cb32-4cc6-b11d-9cba643e3b7b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.282294] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bc3b78a-5e76-4249-ac7a-4739e2f6cc8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.288659] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 824.288659] env[61936]: value = "task-1252854" [ 824.288659] env[61936]: _type = "Task" [ 824.288659] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.295719] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.367066] env[61936]: DEBUG oslo_concurrency.lockutils [req-b5ba6100-069a-43a8-a33f-e4c611e18e6f req-6116e648-963c-491e-b912-9abb8a17c605 service nova] Releasing lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.691940] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 18175943-35d2-4792-a8ea-e54b4ffbaf2f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 824.798264] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503538} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.798535] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] d410e029-cb32-4cc6-b11d-9cba643e3b7b/d410e029-cb32-4cc6-b11d-9cba643e3b7b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.798747] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.798999] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68c95819-e495-4b41-92a4-430458f0568d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.805011] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 824.805011] env[61936]: value = "task-1252855" [ 824.805011] env[61936]: _type = "Task" [ 824.805011] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.812316] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.195917] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance c4f17d38-25a3-4630-8bd4-bb40fa6fddee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 825.315541] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.227787} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.315766] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.316621] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17f7a24-314f-4a53-a322-dc2ca7206a50 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.338820] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] d410e029-cb32-4cc6-b11d-9cba643e3b7b/d410e029-cb32-4cc6-b11d-9cba643e3b7b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.339446] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c72e951e-1410-4d11-9263-898b3287a296 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.357746] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 825.357746] env[61936]: value = "task-1252856" [ 825.357746] env[61936]: _type = "Task" [ 825.357746] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.365275] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.698847] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 258fed0b-ba19-4c90-981a-96bb36ba8bc8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 825.867906] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252856, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.201909] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance aa37292e-8499-46aa-8021-cb8d88b5a35a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 826.368220] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252856, 'name': ReconfigVM_Task, 'duration_secs': 0.734565} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.368477] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Reconfigured VM instance instance-00000036 to attach disk [datastore1] d410e029-cb32-4cc6-b11d-9cba643e3b7b/d410e029-cb32-4cc6-b11d-9cba643e3b7b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.369121] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17ef2b49-0dd0-4127-9157-3132131e0cca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.375506] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 826.375506] env[61936]: value = "task-1252857" [ 826.375506] env[61936]: _type = "Task" [ 826.375506] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.383950] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252857, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.705482] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance efdce04f-b0aa-49a6-94b8-1c960ac33fea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 826.886082] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252857, 'name': Rename_Task, 'duration_secs': 0.139227} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.886383] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.886625] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17ab07cc-c75c-489a-8014-10ce8a8541f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.892604] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 826.892604] env[61936]: value = "task-1252858" [ 826.892604] env[61936]: _type = "Task" [ 826.892604] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.900172] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252858, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.208297] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 827.208604] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 827.208737] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 827.402475] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252858, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.618262] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8996f1-dfab-4de9-b93a-f1ab02483642 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.626144] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af79e04-931f-4922-a159-21d2994ec6c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.656065] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cff7b73-c833-46c9-99f1-6791b0c05a15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.663702] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a223401-d006-49ce-976c-84c143e32ec0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.678060] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.903552] env[61936]: DEBUG oslo_vmware.api [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252858, 'name': PowerOnVM_Task, 'duration_secs': 0.714474} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.903865] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.903865] env[61936]: INFO nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Took 9.28 seconds to spawn the instance on the hypervisor. [ 827.904016] env[61936]: DEBUG nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 827.904938] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a53a40-779f-49f7-9c24-1ede3aa72c41 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.180572] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 828.424020] env[61936]: INFO nova.compute.manager [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Took 29.43 seconds to build instance. [ 828.685817] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 828.686203] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.070s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.686470] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.921s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.688046] env[61936]: INFO nova.compute.claims [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.690713] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 828.690868] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Cleaning up deleted instances {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11626}} [ 828.925995] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b56e8615-e3b8-447b-9cf5-8e435daaf845 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.316s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.116309] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.116475] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.116653] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.116832] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.117130] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.119357] env[61936]: INFO nova.compute.manager [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Terminating instance [ 829.196224] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] There are 4 instances to clean {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11635}} [ 829.196815] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 498a77d6-ca0c-4841-b4a2-b7a024281c6a] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 829.437810] env[61936]: DEBUG nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 829.625689] env[61936]: DEBUG nova.compute.manager [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 829.625871] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.627055] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72fb066-a2f6-4564-be87-6fc3fd0eed68 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.637577] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 829.637577] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4205de6b-96a9-4528-bb03-fbf792d61616 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.643559] env[61936]: DEBUG oslo_vmware.api [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 829.643559] env[61936]: value = "task-1252859" [ 829.643559] env[61936]: _type = "Task" [ 829.643559] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.651625] env[61936]: DEBUG oslo_vmware.api [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252859, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.702512] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 1e0dade4-ea54-457d-8398-54ba810a83c2] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 829.958797] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.012392] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adbcd0d-90bf-4f71-9ac5-3f19085d98c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.020024] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dea795b-d8ad-4ece-9e30-dc6fd2d453cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.048804] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b3d793-c3d7-46af-91d8-5c335205f44b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.056341] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00720782-325e-4e2d-ac0e-bc7af4acde54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.069160] env[61936]: DEBUG nova.compute.provider_tree [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.154274] env[61936]: DEBUG oslo_vmware.api [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252859, 'name': PowerOffVM_Task, 'duration_secs': 0.212629} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.154537] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.154704] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 830.154946] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e11ee0b4-11b0-4a10-a332-dd4088202bfb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.208316] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 994ff54f-a26b-40e0-a2c7-502f3598d9bf] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 830.215601] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.215816] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.216062] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleting the datastore file [datastore1] d410e029-cb32-4cc6-b11d-9cba643e3b7b {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.216325] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09db1a30-c1de-4857-aa6c-115d3e8a5e0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.222437] env[61936]: DEBUG oslo_vmware.api [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 830.222437] env[61936]: value = "task-1252861" [ 830.222437] env[61936]: _type = "Task" [ 830.222437] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.231438] env[61936]: DEBUG oslo_vmware.api [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252861, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.573020] env[61936]: DEBUG nova.scheduler.client.report [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 830.711828] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 7dadef86-7685-40e2-9e68-debc4b4a8a95] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 830.732861] env[61936]: DEBUG oslo_vmware.api [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1252861, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150251} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.733128] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.733307] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 830.733474] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 830.733636] env[61936]: INFO nova.compute.manager [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 830.733981] env[61936]: DEBUG oslo.service.loopingcall [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.734200] env[61936]: DEBUG nova.compute.manager [-] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 830.734320] env[61936]: DEBUG nova.network.neutron [-] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 830.998275] env[61936]: DEBUG nova.compute.manager [req-063ed102-e91c-49a6-87c0-86f1a112ee1d req-ad66ebff-bec9-4398-b092-baf8007d96c3 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Received event network-vif-deleted-7553572a-bf60-44ff-a283-6ab783e30788 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 830.998373] env[61936]: INFO nova.compute.manager [req-063ed102-e91c-49a6-87c0-86f1a112ee1d req-ad66ebff-bec9-4398-b092-baf8007d96c3 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Neutron deleted interface 7553572a-bf60-44ff-a283-6ab783e30788; detaching it from the instance and deleting it from the info cache [ 830.998599] env[61936]: DEBUG nova.network.neutron [req-063ed102-e91c-49a6-87c0-86f1a112ee1d req-ad66ebff-bec9-4398-b092-baf8007d96c3 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.077244] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.077772] env[61936]: DEBUG nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 831.080406] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.601s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.081797] env[61936]: INFO nova.compute.claims [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.215265] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 831.216328] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Cleaning up deleted instances with incomplete migration {{(pid=61936) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11664}} [ 831.453416] env[61936]: DEBUG nova.network.neutron [-] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.501025] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cedc4dde-efd8-411d-a02f-caf5c7104dd1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.511368] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a26185a-9385-4d5c-93c7-48c1baaeff41 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.533745] env[61936]: DEBUG nova.compute.manager [req-063ed102-e91c-49a6-87c0-86f1a112ee1d req-ad66ebff-bec9-4398-b092-baf8007d96c3 service nova] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Detach interface failed, port_id=7553572a-bf60-44ff-a283-6ab783e30788, reason: Instance d410e029-cb32-4cc6-b11d-9cba643e3b7b could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 831.586430] env[61936]: DEBUG nova.compute.utils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.589752] env[61936]: DEBUG nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 831.589887] env[61936]: DEBUG nova.network.neutron [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.718587] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 831.890028] env[61936]: DEBUG nova.policy [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a346002c214e4b238f323ffffc6b61b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c334e9711bc64304ab017eb6ab3e2d23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.956247] env[61936]: INFO nova.compute.manager [-] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Took 1.22 seconds to deallocate network for instance. [ 832.090392] env[61936]: DEBUG nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 832.202746] env[61936]: DEBUG nova.network.neutron [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Successfully created port: 1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.420208] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225cf01e-a893-4277-a914-6a017b90dc9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.428088] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad4b72c-83f1-40cf-b198-d7e4213c9d43 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.460945] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28453fc8-4ffb-472e-8206-fd298a8531be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.463604] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.469239] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6404cada-e55d-4f02-b9d0-60bf07080642 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.482963] env[61936]: DEBUG nova.compute.provider_tree [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.986609] env[61936]: DEBUG nova.scheduler.client.report [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 833.107147] env[61936]: DEBUG nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 833.131390] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 833.131724] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.131926] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 833.132131] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.132278] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 833.132422] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 833.132622] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 833.132777] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 833.132936] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 833.134197] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 833.134900] env[61936]: DEBUG nova.virt.hardware [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 833.136103] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83372538-6de8-475f-ba86-3e9dabd67e34 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.145653] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb02221-1972-4be7-861e-f915a5e85fab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.491845] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.492412] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 833.495039] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.079s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.496427] env[61936]: INFO nova.compute.claims [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.722629] env[61936]: DEBUG nova.compute.manager [req-ace16efb-b0a7-436f-a548-4aca215a9208 req-88ee6253-cd8c-4e2d-baf5-28a0250d95b9 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received event network-vif-plugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 833.722863] env[61936]: DEBUG oslo_concurrency.lockutils [req-ace16efb-b0a7-436f-a548-4aca215a9208 req-88ee6253-cd8c-4e2d-baf5-28a0250d95b9 service nova] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.723107] env[61936]: DEBUG oslo_concurrency.lockutils [req-ace16efb-b0a7-436f-a548-4aca215a9208 req-88ee6253-cd8c-4e2d-baf5-28a0250d95b9 service nova] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.723300] env[61936]: DEBUG oslo_concurrency.lockutils [req-ace16efb-b0a7-436f-a548-4aca215a9208 req-88ee6253-cd8c-4e2d-baf5-28a0250d95b9 service nova] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.726014] env[61936]: DEBUG nova.compute.manager [req-ace16efb-b0a7-436f-a548-4aca215a9208 req-88ee6253-cd8c-4e2d-baf5-28a0250d95b9 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] No waiting events found dispatching network-vif-plugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 833.726014] env[61936]: WARNING nova.compute.manager [req-ace16efb-b0a7-436f-a548-4aca215a9208 req-88ee6253-cd8c-4e2d-baf5-28a0250d95b9 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received unexpected event network-vif-plugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 for instance with vm_state building and task_state spawning. [ 833.823383] env[61936]: DEBUG nova.network.neutron [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Successfully updated port: 1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.000915] env[61936]: DEBUG nova.compute.utils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.004265] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 834.004436] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 834.042308] env[61936]: DEBUG nova.policy [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4132cf491b042a68352893e929ca439', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c586842b2cf34c1e999dc110ba05261b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.327430] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.327430] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.327430] env[61936]: DEBUG nova.network.neutron [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.422994] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Successfully created port: 74d05192-a7d3-48d2-b0e1-62141112c67e {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.505101] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 834.856968] env[61936]: DEBUG nova.network.neutron [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.894032] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839d1387-c253-48d8-9042-046868a023fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.901418] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b502429b-7f85-4814-98fb-4157275237f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.936463] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62257deb-a395-47d8-9107-ca16f68a0ec9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.944016] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d744f23-4bc6-4e3b-9901-c527bf4276e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.956999] env[61936]: DEBUG nova.compute.provider_tree [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.038574] env[61936]: DEBUG nova.network.neutron [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.463060] env[61936]: DEBUG nova.scheduler.client.report [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 835.516073] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 835.540781] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 835.541038] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.541202] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 835.541381] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.541524] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 835.541667] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 835.541864] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 835.542028] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 835.542195] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 835.542351] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 835.542615] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 835.542981] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.543275] env[61936]: DEBUG nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Instance network_info: |[{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 835.544354] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dedbe3-f73f-487b-9949-23c9e5d055e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.547117] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:fd:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a44713-0af1-486e-bc0d-00e03a769fa4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1cdd9a15-01e7-40cf-80fc-24bea0b642d3', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.554426] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating folder: Project (c334e9711bc64304ab017eb6ab3e2d23). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.554665] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2c9ba80-48eb-48c8-b326-97b8a305d7a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.562026] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0453ae-eecb-420a-996e-710a96a2ea3b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.566447] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Created folder: Project (c334e9711bc64304ab017eb6ab3e2d23) in parent group-v269874. [ 835.566619] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating folder: Instances. Parent ref: group-v269916. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.567125] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df0d9dff-f43d-4bc0-b92d-13f98bfd0144 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.583139] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Created folder: Instances in parent group-v269916. [ 835.583351] env[61936]: DEBUG oslo.service.loopingcall [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.583521] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.583705] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38f62346-3b14-4665-bb93-177c9200be05 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.602259] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.602259] env[61936]: value = "task-1252864" [ 835.602259] env[61936]: _type = "Task" [ 835.602259] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.609479] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252864, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.753012] env[61936]: DEBUG nova.compute.manager [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received event network-changed-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 835.753271] env[61936]: DEBUG nova.compute.manager [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Refreshing instance network info cache due to event network-changed-1cdd9a15-01e7-40cf-80fc-24bea0b642d3. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 835.753504] env[61936]: DEBUG oslo_concurrency.lockutils [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.753631] env[61936]: DEBUG oslo_concurrency.lockutils [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.753791] env[61936]: DEBUG nova.network.neutron [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Refreshing network info cache for port 1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.967151] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.967654] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 835.970306] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.271s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.972760] env[61936]: INFO nova.compute.claims [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.014667] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Successfully updated port: 74d05192-a7d3-48d2-b0e1-62141112c67e {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.111909] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252864, 'name': CreateVM_Task, 'duration_secs': 0.293832} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.112225] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.112754] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.112918] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.113328] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.113482] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4737e19-85da-4cae-8ed2-e7a0a3ef4394 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.117663] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 836.117663] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5299a104-0a64-a638-c32f-862f2874ce3e" [ 836.117663] env[61936]: _type = "Task" [ 836.117663] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.125182] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5299a104-0a64-a638-c32f-862f2874ce3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.482979] env[61936]: DEBUG nova.compute.utils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.486533] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 836.486701] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 836.520534] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "refresh_cache-bdea3742-5789-4e97-b139-33cfee134843" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.520691] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "refresh_cache-bdea3742-5789-4e97-b139-33cfee134843" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.520845] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.529653] env[61936]: DEBUG nova.policy [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4132cf491b042a68352893e929ca439', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c586842b2cf34c1e999dc110ba05261b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.561169] env[61936]: DEBUG nova.network.neutron [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updated VIF entry in instance network info cache for port 1cdd9a15-01e7-40cf-80fc-24bea0b642d3. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.561537] env[61936]: DEBUG nova.network.neutron [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.628940] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5299a104-0a64-a638-c32f-862f2874ce3e, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.629195] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.629424] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.629655] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.629796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.629974] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.630254] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dc9fef4-2ca8-4008-b602-023bdc930dd6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.638465] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.638636] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.639377] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdcca076-41b2-4029-9638-0bc1be558e99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.644169] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 836.644169] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e21514-2ccf-b02e-cbb6-d85619f5f7af" [ 836.644169] env[61936]: _type = "Task" [ 836.644169] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.651854] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e21514-2ccf-b02e-cbb6-d85619f5f7af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.934130] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Successfully created port: a5a09509-5534-4354-b48f-c09cfa35b3d6 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.987338] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 837.070518] env[61936]: DEBUG oslo_concurrency.lockutils [req-26bcd6cc-a9f5-458a-9e29-13f3c37ba121 req-26d08182-915d-49b7-b6ba-23f1077ed7aa service nova] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.071818] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.160062] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e21514-2ccf-b02e-cbb6-d85619f5f7af, 'name': SearchDatastore_Task, 'duration_secs': 0.007908} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.160894] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49901adb-46d4-45a6-ae52-8581ff7aa272 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.168451] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 837.168451] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b5abc9-e4a6-3c98-8cc5-dc24a898d43c" [ 837.168451] env[61936]: _type = "Task" [ 837.168451] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.179031] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b5abc9-e4a6-3c98-8cc5-dc24a898d43c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.233222] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Updating instance_info_cache with network_info: [{"id": "74d05192-a7d3-48d2-b0e1-62141112c67e", "address": "fa:16:3e:33:ab:1b", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74d05192-a7", "ovs_interfaceid": "74d05192-a7d3-48d2-b0e1-62141112c67e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.360538] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b56640d-3e57-4e37-8f10-92f1ccb2f36c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.368162] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2168f4c2-5c66-4052-927e-a38f711389f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.398640] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0df563-f050-4135-a830-6211b1336ca4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.405954] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24be2ce6-de37-4032-b710-1374aa6289f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.418860] env[61936]: DEBUG nova.compute.provider_tree [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.678826] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b5abc9-e4a6-3c98-8cc5-dc24a898d43c, 'name': SearchDatastore_Task, 'duration_secs': 0.008746} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.678826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.678826] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.679131] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af9c0020-e644-4c75-bd92-9ba59342ecdb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.685054] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 837.685054] env[61936]: value = "task-1252865" [ 837.685054] env[61936]: _type = "Task" [ 837.685054] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.692180] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252865, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.739614] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "refresh_cache-bdea3742-5789-4e97-b139-33cfee134843" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.739614] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Instance network_info: |[{"id": "74d05192-a7d3-48d2-b0e1-62141112c67e", "address": "fa:16:3e:33:ab:1b", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74d05192-a7", "ovs_interfaceid": "74d05192-a7d3-48d2-b0e1-62141112c67e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 837.739792] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:ab:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '74d05192-a7d3-48d2-b0e1-62141112c67e', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.747582] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Creating folder: Project (c586842b2cf34c1e999dc110ba05261b). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.747910] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-313dd8d4-d2de-4f86-b830-3bf761337ce8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.758891] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Created folder: Project (c586842b2cf34c1e999dc110ba05261b) in parent group-v269874. [ 837.758891] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Creating folder: Instances. Parent ref: group-v269919. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.759174] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2ffa8b3-4dc0-4135-a894-7fbe8d32a7a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.767453] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Created folder: Instances in parent group-v269919. [ 837.767743] env[61936]: DEBUG oslo.service.loopingcall [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.767965] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdea3742-5789-4e97-b139-33cfee134843] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.768227] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75eb4a97-9f05-4d1e-a6f5-d5886608c08f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.790177] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.790177] env[61936]: value = "task-1252868" [ 837.790177] env[61936]: _type = "Task" [ 837.790177] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.800077] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.802421] env[61936]: DEBUG nova.compute.manager [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Received event network-vif-plugged-74d05192-a7d3-48d2-b0e1-62141112c67e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 837.802590] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] Acquiring lock "bdea3742-5789-4e97-b139-33cfee134843-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.802800] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] Lock "bdea3742-5789-4e97-b139-33cfee134843-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.802995] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] Lock "bdea3742-5789-4e97-b139-33cfee134843-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.803188] env[61936]: DEBUG nova.compute.manager [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] No waiting events found dispatching network-vif-plugged-74d05192-a7d3-48d2-b0e1-62141112c67e {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 837.803354] env[61936]: WARNING nova.compute.manager [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Received unexpected event network-vif-plugged-74d05192-a7d3-48d2-b0e1-62141112c67e for instance with vm_state building and task_state spawning. [ 837.803508] env[61936]: DEBUG nova.compute.manager [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Received event network-changed-74d05192-a7d3-48d2-b0e1-62141112c67e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 837.803654] env[61936]: DEBUG nova.compute.manager [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Refreshing instance network info cache due to event network-changed-74d05192-a7d3-48d2-b0e1-62141112c67e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 837.803829] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] Acquiring lock "refresh_cache-bdea3742-5789-4e97-b139-33cfee134843" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.803957] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] Acquired lock "refresh_cache-bdea3742-5789-4e97-b139-33cfee134843" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.804120] env[61936]: DEBUG nova.network.neutron [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Refreshing network info cache for port 74d05192-a7d3-48d2-b0e1-62141112c67e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.922142] env[61936]: DEBUG nova.scheduler.client.report [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 837.999701] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 838.027170] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 838.027468] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.027642] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 838.027829] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.027975] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 838.028166] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 838.028385] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 838.028591] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 838.028761] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 838.028923] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 838.029124] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 838.030200] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b086165-2299-49ea-85f3-e3b8e5dc43ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.039158] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76100bde-05bd-47fe-abaf-86441f584657 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.194933] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252865, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42847} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.195257] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.195428] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.195682] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-583cec0f-da0b-47b0-8631-28032e594d14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.202693] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 838.202693] env[61936]: value = "task-1252869" [ 838.202693] env[61936]: _type = "Task" [ 838.202693] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.210644] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252869, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.305096] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.343999] env[61936]: DEBUG nova.compute.manager [req-85826402-1885-4830-8bab-b9eae502ffd6 req-2ebc1517-f425-4fec-b78e-2a53b284adaa service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Received event network-vif-plugged-a5a09509-5534-4354-b48f-c09cfa35b3d6 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 838.344242] env[61936]: DEBUG oslo_concurrency.lockutils [req-85826402-1885-4830-8bab-b9eae502ffd6 req-2ebc1517-f425-4fec-b78e-2a53b284adaa service nova] Acquiring lock "77a59650-d28b-4862-bb35-6910e6f7d6c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.344443] env[61936]: DEBUG oslo_concurrency.lockutils [req-85826402-1885-4830-8bab-b9eae502ffd6 req-2ebc1517-f425-4fec-b78e-2a53b284adaa service nova] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.344605] env[61936]: DEBUG oslo_concurrency.lockutils [req-85826402-1885-4830-8bab-b9eae502ffd6 req-2ebc1517-f425-4fec-b78e-2a53b284adaa service nova] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.344830] env[61936]: DEBUG nova.compute.manager [req-85826402-1885-4830-8bab-b9eae502ffd6 req-2ebc1517-f425-4fec-b78e-2a53b284adaa service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] No waiting events found dispatching network-vif-plugged-a5a09509-5534-4354-b48f-c09cfa35b3d6 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 838.344914] env[61936]: WARNING nova.compute.manager [req-85826402-1885-4830-8bab-b9eae502ffd6 req-2ebc1517-f425-4fec-b78e-2a53b284adaa service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Received unexpected event network-vif-plugged-a5a09509-5534-4354-b48f-c09cfa35b3d6 for instance with vm_state building and task_state spawning. [ 838.428454] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.428966] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 838.432382] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.812s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.436339] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.004s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.438519] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.218s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.443399] env[61936]: INFO nova.compute.claims [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.454736] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Successfully updated port: a5a09509-5534-4354-b48f-c09cfa35b3d6 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.474458] env[61936]: INFO nova.scheduler.client.report [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Deleted allocations for instance 1251e7f5-684c-428f-9d63-60b77084d09e [ 838.569823] env[61936]: DEBUG nova.network.neutron [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Updated VIF entry in instance network info cache for port 74d05192-a7d3-48d2-b0e1-62141112c67e. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.570224] env[61936]: DEBUG nova.network.neutron [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Updating instance_info_cache with network_info: [{"id": "74d05192-a7d3-48d2-b0e1-62141112c67e", "address": "fa:16:3e:33:ab:1b", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap74d05192-a7", "ovs_interfaceid": "74d05192-a7d3-48d2-b0e1-62141112c67e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.712978] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252869, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.203529} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.713276] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.714073] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d02d47-c269-44e2-9cde-facaa58faee0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.740426] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.740719] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ba27693-c42f-4cff-b927-1101381252d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.760687] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 838.760687] env[61936]: value = "task-1252870" [ 838.760687] env[61936]: _type = "Task" [ 838.760687] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.768311] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252870, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.800646] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.949576] env[61936]: DEBUG nova.compute.utils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.951073] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 838.951269] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.958375] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "refresh_cache-77a59650-d28b-4862-bb35-6910e6f7d6c9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.958375] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "refresh_cache-77a59650-d28b-4862-bb35-6910e6f7d6c9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.958375] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 838.984807] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea9fdf60-1db6-46b7-bfcf-f48445faf666 tempest-ServerAddressesTestJSON-206545344 tempest-ServerAddressesTestJSON-206545344-project-member] Lock "1251e7f5-684c-428f-9d63-60b77084d09e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.744s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.040505] env[61936]: DEBUG nova.policy [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4132cf491b042a68352893e929ca439', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c586842b2cf34c1e999dc110ba05261b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.072594] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ec6fe5f-2fac-4579-9adc-a3f48074cdca req-ae51cc11-d410-4e2f-86e4-a83c85978eda service nova] Releasing lock "refresh_cache-bdea3742-5789-4e97-b139-33cfee134843" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.271262] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252870, 'name': ReconfigVM_Task, 'duration_secs': 0.285916} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.271574] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.272169] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e0395c9-d209-4991-a0e4-e5d192919830 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.282047] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 839.282047] env[61936]: value = "task-1252871" [ 839.282047] env[61936]: _type = "Task" [ 839.282047] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.287911] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252871, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.301199] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.454628] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 839.509112] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.687169] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Updating instance_info_cache with network_info: [{"id": "a5a09509-5534-4354-b48f-c09cfa35b3d6", "address": "fa:16:3e:cf:fe:9e", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5a09509-55", "ovs_interfaceid": "a5a09509-5534-4354-b48f-c09cfa35b3d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.789218] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Successfully created port: 853ddccf-d49b-471f-afc6-5cf28851a036 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.809619] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252871, 'name': Rename_Task, 'duration_secs': 0.222839} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.811898] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.814103] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47b98d2b-dd5a-4e7d-9cf0-fc67dae550f0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.825155] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.830980] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6409609-6fda-401e-a5dd-a0e95e8ed3fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.839024] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 839.839024] env[61936]: value = "task-1252872" [ 839.839024] env[61936]: _type = "Task" [ 839.839024] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.847603] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dcf425b-1fee-4e75-960b-1a16a7b3a187 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.858347] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252872, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.907649] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574c7c27-bfc3-49f1-9697-c8e70e3e97ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.918649] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670a187b-50b9-4ee5-a303-f0201b48a874 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.941523] env[61936]: DEBUG nova.compute.provider_tree [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.190723] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "refresh_cache-77a59650-d28b-4862-bb35-6910e6f7d6c9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.191111] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Instance network_info: |[{"id": "a5a09509-5534-4354-b48f-c09cfa35b3d6", "address": "fa:16:3e:cf:fe:9e", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5a09509-55", "ovs_interfaceid": "a5a09509-5534-4354-b48f-c09cfa35b3d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 840.191569] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:fe:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5a09509-5534-4354-b48f-c09cfa35b3d6', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.199856] env[61936]: DEBUG oslo.service.loopingcall [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.200305] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 840.200711] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46d59b2b-bea9-4fb2-b037-64721914b99f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.224235] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.224235] env[61936]: value = "task-1252873" [ 840.224235] env[61936]: _type = "Task" [ 840.224235] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.232305] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252873, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.307451] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.348555] env[61936]: DEBUG oslo_vmware.api [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1252872, 'name': PowerOnVM_Task, 'duration_secs': 0.494125} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.348702] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.348785] env[61936]: INFO nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Took 7.24 seconds to spawn the instance on the hypervisor. [ 840.348955] env[61936]: DEBUG nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 840.350088] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2475afe7-2b0c-48aa-8afd-7bdb7330b9ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.380933] env[61936]: DEBUG nova.compute.manager [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Received event network-changed-a5a09509-5534-4354-b48f-c09cfa35b3d6 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 840.381725] env[61936]: DEBUG nova.compute.manager [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Refreshing instance network info cache due to event network-changed-a5a09509-5534-4354-b48f-c09cfa35b3d6. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 840.381994] env[61936]: DEBUG oslo_concurrency.lockutils [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] Acquiring lock "refresh_cache-77a59650-d28b-4862-bb35-6910e6f7d6c9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.382197] env[61936]: DEBUG oslo_concurrency.lockutils [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] Acquired lock "refresh_cache-77a59650-d28b-4862-bb35-6910e6f7d6c9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.382682] env[61936]: DEBUG nova.network.neutron [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Refreshing network info cache for port a5a09509-5534-4354-b48f-c09cfa35b3d6 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.448842] env[61936]: DEBUG nova.scheduler.client.report [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 840.467780] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 840.509208] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 840.509208] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.509208] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 840.509356] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.509356] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 840.509356] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 840.509356] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 840.509356] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 840.509531] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 840.509564] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 840.509939] env[61936]: DEBUG nova.virt.hardware [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 840.511901] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3446cb08-0937-46b8-97dc-72b1da04a092 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.526445] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44034f4d-6cf6-46d2-ac99-d062e856d532 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.735028] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252873, 'name': CreateVM_Task, 'duration_secs': 0.371337} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.735215] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 840.735923] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.736107] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.736443] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.736703] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eac61c9-d846-499e-9573-4ba0aeec3481 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.742024] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 840.742024] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239acef-08a2-89a0-471c-669e570d2fb1" [ 840.742024] env[61936]: _type = "Task" [ 840.742024] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.752515] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239acef-08a2-89a0-471c-669e570d2fb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.806115] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.869382] env[61936]: INFO nova.compute.manager [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Took 37.12 seconds to build instance. [ 840.955382] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.955965] env[61936]: DEBUG nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 840.961592] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.824s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.961868] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.963843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.636s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.965287] env[61936]: INFO nova.compute.claims [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.991766] env[61936]: INFO nova.scheduler.client.report [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted allocations for instance 144ddda8-db07-4308-8fd3-4045067b5b37 [ 841.142352] env[61936]: DEBUG nova.network.neutron [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Updated VIF entry in instance network info cache for port a5a09509-5534-4354-b48f-c09cfa35b3d6. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 841.142717] env[61936]: DEBUG nova.network.neutron [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Updating instance_info_cache with network_info: [{"id": "a5a09509-5534-4354-b48f-c09cfa35b3d6", "address": "fa:16:3e:cf:fe:9e", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5a09509-55", "ovs_interfaceid": "a5a09509-5534-4354-b48f-c09cfa35b3d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.255792] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239acef-08a2-89a0-471c-669e570d2fb1, 'name': SearchDatastore_Task, 'duration_secs': 0.014034} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.256115] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.256369] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.256615] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.256860] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.256928] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.261741] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e61a26f-7565-4330-8e55-2346cd4ea709 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.268036] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.268306] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.269094] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d05c53c9-2d29-4f4f-be67-74b6880c03a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.274994] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 841.274994] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52751800-afce-9ae4-b729-a014ddfb976f" [ 841.274994] env[61936]: _type = "Task" [ 841.274994] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.283310] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52751800-afce-9ae4-b729-a014ddfb976f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.306977] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252868, 'name': CreateVM_Task, 'duration_secs': 3.479123} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.307295] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdea3742-5789-4e97-b139-33cfee134843] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 841.307889] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.308166] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.308624] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.310195] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b53bbad-90dc-4fc8-af63-569ce6b337b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.314394] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 841.314394] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521e2757-1622-6a2e-bc66-3d0633b68b83" [ 841.314394] env[61936]: _type = "Task" [ 841.314394] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.325316] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521e2757-1622-6a2e-bc66-3d0633b68b83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.372655] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0bf4452-4483-45c2-8268-79c5152f921a tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.704s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.473622] env[61936]: DEBUG nova.compute.utils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.478035] env[61936]: DEBUG nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Not allocating networking since 'none' was specified. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 841.499787] env[61936]: DEBUG oslo_concurrency.lockutils [None req-daca813f-a60e-4359-8d5a-6a1d3e840102 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "144ddda8-db07-4308-8fd3-4045067b5b37" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.799s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.645348] env[61936]: DEBUG oslo_concurrency.lockutils [req-c9f8392a-1429-412a-bb55-fdc9cb8c6ebb req-3ef1aa74-b209-4c36-9b2e-73d68ec7d684 service nova] Releasing lock "refresh_cache-77a59650-d28b-4862-bb35-6910e6f7d6c9" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.785936] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52751800-afce-9ae4-b729-a014ddfb976f, 'name': SearchDatastore_Task, 'duration_secs': 0.012008} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.786988] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08f574d5-924a-49b6-9954-f1c576832e02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.792562] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 841.792562] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522871cd-e8e7-56a8-9ed3-880e73e456b8" [ 841.792562] env[61936]: _type = "Task" [ 841.792562] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.801937] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522871cd-e8e7-56a8-9ed3-880e73e456b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.824658] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521e2757-1622-6a2e-bc66-3d0633b68b83, 'name': SearchDatastore_Task, 'duration_secs': 0.012009} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.825213] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.825490] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.825895] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.836928] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Successfully updated port: 853ddccf-d49b-471f-afc6-5cf28851a036 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.878648] env[61936]: DEBUG nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 841.925846] env[61936]: DEBUG nova.compute.manager [req-555e9184-118f-4a85-b01a-e734da701ab8 req-aaf978b6-5e94-42e2-b23f-2cb16e5ba9c0 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Received event network-vif-plugged-853ddccf-d49b-471f-afc6-5cf28851a036 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 841.929621] env[61936]: DEBUG oslo_concurrency.lockutils [req-555e9184-118f-4a85-b01a-e734da701ab8 req-aaf978b6-5e94-42e2-b23f-2cb16e5ba9c0 service nova] Acquiring lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.929921] env[61936]: DEBUG oslo_concurrency.lockutils [req-555e9184-118f-4a85-b01a-e734da701ab8 req-aaf978b6-5e94-42e2-b23f-2cb16e5ba9c0 service nova] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.004s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.930112] env[61936]: DEBUG oslo_concurrency.lockutils [req-555e9184-118f-4a85-b01a-e734da701ab8 req-aaf978b6-5e94-42e2-b23f-2cb16e5ba9c0 service nova] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.930298] env[61936]: DEBUG nova.compute.manager [req-555e9184-118f-4a85-b01a-e734da701ab8 req-aaf978b6-5e94-42e2-b23f-2cb16e5ba9c0 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] No waiting events found dispatching network-vif-plugged-853ddccf-d49b-471f-afc6-5cf28851a036 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 841.930464] env[61936]: WARNING nova.compute.manager [req-555e9184-118f-4a85-b01a-e734da701ab8 req-aaf978b6-5e94-42e2-b23f-2cb16e5ba9c0 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Received unexpected event network-vif-plugged-853ddccf-d49b-471f-afc6-5cf28851a036 for instance with vm_state building and task_state spawning. [ 841.983033] env[61936]: DEBUG nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 842.264453] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48bae5b-ee7f-4d13-9bec-8bb7b2a600fe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.272396] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628b7bea-da50-49ba-ac72-3c990e345067 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.306284] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3378e9e-49e6-4066-9b7f-b99363bbc4e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.313802] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522871cd-e8e7-56a8-9ed3-880e73e456b8, 'name': SearchDatastore_Task, 'duration_secs': 0.009814} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.315775] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.316054] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 77a59650-d28b-4862-bb35-6910e6f7d6c9/77a59650-d28b-4862-bb35-6910e6f7d6c9.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.316373] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.316559] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.316755] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70cf1cef-faf8-4680-85e0-28b82f1a6b2d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.319402] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb45e9c-c550-48d9-9650-d5d6797867e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.323072] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a16e7ca-530e-4eff-8b10-7430eacd4cc0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.335029] env[61936]: DEBUG nova.compute.provider_tree [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.337912] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.338196] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 842.339394] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 842.339394] env[61936]: value = "task-1252874" [ 842.339394] env[61936]: _type = "Task" [ 842.339394] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.339825] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f6f1f31-9311-4f4c-b677-52c5fc4edd89 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.342399] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "refresh_cache-fafc995e-4c93-4ca0-b078-24a1ae0ab427" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.342525] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "refresh_cache-fafc995e-4c93-4ca0-b078-24a1ae0ab427" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.342663] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.350218] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 842.350218] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5233f158-d0e8-7135-2e79-6ff3b06b0f87" [ 842.350218] env[61936]: _type = "Task" [ 842.350218] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.353749] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252874, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.362356] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5233f158-d0e8-7135-2e79-6ff3b06b0f87, 'name': SearchDatastore_Task, 'duration_secs': 0.008136} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.362910] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14ca8a87-90cd-4dd6-b6a5-cc7c5eed998c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.369027] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 842.369027] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5253155b-d25b-7ef8-46e5-4c7267a766dd" [ 842.369027] env[61936]: _type = "Task" [ 842.369027] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.376089] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5253155b-d25b-7ef8-46e5-4c7267a766dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.383412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.383638] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.401393] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.419618] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "628882ff-2f0f-49ae-8454-d3791fb55545" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.419833] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.841095] env[61936]: DEBUG nova.scheduler.client.report [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 842.855036] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252874, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.40712} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.855298] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 77a59650-d28b-4862-bb35-6910e6f7d6c9/77a59650-d28b-4862-bb35-6910e6f7d6c9.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 842.856081] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 842.856081] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b507a04-82a6-470b-aabe-d326a0710ac3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.862740] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 842.862740] env[61936]: value = "task-1252875" [ 842.862740] env[61936]: _type = "Task" [ 842.862740] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.873499] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252875, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.881810] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5253155b-d25b-7ef8-46e5-4c7267a766dd, 'name': SearchDatastore_Task, 'duration_secs': 0.007664} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.882532] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.884583] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.884884] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] bdea3742-5789-4e97-b139-33cfee134843/bdea3742-5789-4e97-b139-33cfee134843.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.886291] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4952381-29a2-45df-9f07-13fd6c1307fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.892819] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 842.892819] env[61936]: value = "task-1252876" [ 842.892819] env[61936]: _type = "Task" [ 842.892819] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.901696] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.992027] env[61936]: DEBUG nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 843.017957] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 843.018231] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.018411] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 843.018560] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.018703] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 843.018844] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 843.019080] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 843.019247] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 843.019415] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 843.019575] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 843.019743] env[61936]: DEBUG nova.virt.hardware [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 843.020967] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e390a2b1-2824-46f1-ad0b-8ccd33660658 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.029291] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4db6a21-3e68-4320-b2ff-eed1b01fd291 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.042849] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.048657] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Creating folder: Project (365a03e2cf0744ccbd9b55dfd072eb96). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.051210] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6583a39f-2b2e-4c22-b6aa-9570386fcdf1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.060913] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Created folder: Project (365a03e2cf0744ccbd9b55dfd072eb96) in parent group-v269874. [ 843.061165] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Creating folder: Instances. Parent ref: group-v269923. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.061442] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34e580fd-f005-448b-8b06-851920724b2c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.071519] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Created folder: Instances in parent group-v269923. [ 843.071786] env[61936]: DEBUG oslo.service.loopingcall [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.071995] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.072255] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9ec1ee6-180a-497d-988d-cf8047a4c24e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.088704] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.088704] env[61936]: value = "task-1252879" [ 843.088704] env[61936]: _type = "Task" [ 843.088704] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.096711] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252879, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.138311] env[61936]: DEBUG nova.network.neutron [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Updating instance_info_cache with network_info: [{"id": "853ddccf-d49b-471f-afc6-5cf28851a036", "address": "fa:16:3e:fd:9c:aa", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853ddccf-d4", "ovs_interfaceid": "853ddccf-d49b-471f-afc6-5cf28851a036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.349529] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.350083] env[61936]: DEBUG nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 843.352869] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.448s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.354356] env[61936]: INFO nova.compute.claims [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.373836] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252875, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063274} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.374131] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.374971] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c999cc-ed67-4e46-ade2-bd2109baba99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.397691] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 77a59650-d28b-4862-bb35-6910e6f7d6c9/77a59650-d28b-4862-bb35-6910e6f7d6c9.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.399227] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-341efea5-5f67-4fea-8f4a-72aa431c18ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.422795] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524251} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.424250] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] bdea3742-5789-4e97-b139-33cfee134843/bdea3742-5789-4e97-b139-33cfee134843.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 843.424480] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.425157] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 843.425157] env[61936]: value = "task-1252880" [ 843.425157] env[61936]: _type = "Task" [ 843.425157] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.425364] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3eb75fd7-29e3-495f-89f5-a3940592b4ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.437222] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252880, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.438672] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 843.438672] env[61936]: value = "task-1252881" [ 843.438672] env[61936]: _type = "Task" [ 843.438672] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.446089] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252881, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.600225] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252879, 'name': CreateVM_Task, 'duration_secs': 0.47953} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.600398] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.600778] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.600939] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.601326] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.601569] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-363c8801-695e-4147-92ed-6efc5ab9599d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.605721] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 843.605721] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5269ed5b-4ff9-0f9c-7449-580fadb10bbf" [ 843.605721] env[61936]: _type = "Task" [ 843.605721] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.613280] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5269ed5b-4ff9-0f9c-7449-580fadb10bbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.639987] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "refresh_cache-fafc995e-4c93-4ca0-b078-24a1ae0ab427" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.640382] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Instance network_info: |[{"id": "853ddccf-d49b-471f-afc6-5cf28851a036", "address": "fa:16:3e:fd:9c:aa", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853ddccf-d4", "ovs_interfaceid": "853ddccf-d49b-471f-afc6-5cf28851a036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 843.640866] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:9c:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '853ddccf-d49b-471f-afc6-5cf28851a036', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.648183] env[61936]: DEBUG oslo.service.loopingcall [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.648404] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.648623] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9e5b2456-7531-47af-b2d4-bede0f282b5d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.668037] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.668037] env[61936]: value = "task-1252882" [ 843.668037] env[61936]: _type = "Task" [ 843.668037] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.675531] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252882, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.859072] env[61936]: DEBUG nova.compute.utils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.862250] env[61936]: DEBUG nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 843.862417] env[61936]: DEBUG nova.network.neutron [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.902386] env[61936]: DEBUG nova.policy [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b893b7bbf3e4ead9fe4351b5be25eeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a643413e3d7748cbac250b824fcbba39', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.937299] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252880, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.949158] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252881, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.23728} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.949435] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.950231] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecae713-bb7b-447b-aba7-d9b1e9187f01 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.954930] env[61936]: DEBUG nova.compute.manager [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Received event network-changed-853ddccf-d49b-471f-afc6-5cf28851a036 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 843.955130] env[61936]: DEBUG nova.compute.manager [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Refreshing instance network info cache due to event network-changed-853ddccf-d49b-471f-afc6-5cf28851a036. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 843.955339] env[61936]: DEBUG oslo_concurrency.lockutils [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] Acquiring lock "refresh_cache-fafc995e-4c93-4ca0-b078-24a1ae0ab427" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.955478] env[61936]: DEBUG oslo_concurrency.lockutils [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] Acquired lock "refresh_cache-fafc995e-4c93-4ca0-b078-24a1ae0ab427" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.955631] env[61936]: DEBUG nova.network.neutron [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Refreshing network info cache for port 853ddccf-d49b-471f-afc6-5cf28851a036 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 843.978016] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] bdea3742-5789-4e97-b139-33cfee134843/bdea3742-5789-4e97-b139-33cfee134843.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.981465] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3be4aa2d-bb17-4b11-823b-925bb33fd408 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.998991] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 843.998991] env[61936]: value = "task-1252883" [ 843.998991] env[61936]: _type = "Task" [ 843.998991] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.006909] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252883, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.116356] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5269ed5b-4ff9-0f9c-7449-580fadb10bbf, 'name': SearchDatastore_Task, 'duration_secs': 0.021192} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.116937] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.117230] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.117408] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.117513] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.117761] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.118186] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-752bf68e-6718-4c77-9210-be62aa4e62ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.126425] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.126650] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.127559] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29ba0a43-35b9-4d8d-a07a-4c2f5e29f96b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.133040] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 844.133040] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52213b49-deb0-82bd-2c31-7975d5d0248e" [ 844.133040] env[61936]: _type = "Task" [ 844.133040] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.142709] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52213b49-deb0-82bd-2c31-7975d5d0248e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.178594] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252882, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.212712] env[61936]: DEBUG nova.network.neutron [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Successfully created port: ca6d9c8e-18c6-4a80-a792-baa44a0acc54 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.363686] env[61936]: DEBUG nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 844.440635] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252880, 'name': ReconfigVM_Task, 'duration_secs': 0.546812} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.440947] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 77a59650-d28b-4862-bb35-6910e6f7d6c9/77a59650-d28b-4862-bb35-6910e6f7d6c9.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.441766] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cac4317d-8373-4a2e-9aa0-7a8d347b33fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.450430] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 844.450430] env[61936]: value = "task-1252884" [ 844.450430] env[61936]: _type = "Task" [ 844.450430] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.463275] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252884, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.512379] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252883, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.647282] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52213b49-deb0-82bd-2c31-7975d5d0248e, 'name': SearchDatastore_Task, 'duration_secs': 0.010224} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.650054] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-517ffe8a-eb03-445e-aa8d-aa1ff968d70a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.654802] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 844.654802] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5283a374-40b6-ba70-dad1-b612642ee990" [ 844.654802] env[61936]: _type = "Task" [ 844.654802] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.662409] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5283a374-40b6-ba70-dad1-b612642ee990, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.678177] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252882, 'name': CreateVM_Task, 'duration_secs': 0.784011} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.680320] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.682966] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.683116] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.683416] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.684115] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5314fbe9-ae90-4c47-a003-e64f9027f893 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.688306] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 844.688306] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525fd798-fec4-4917-32ae-367f399f66d7" [ 844.688306] env[61936]: _type = "Task" [ 844.688306] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.697891] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525fd798-fec4-4917-32ae-367f399f66d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.734645] env[61936]: DEBUG nova.network.neutron [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Updated VIF entry in instance network info cache for port 853ddccf-d49b-471f-afc6-5cf28851a036. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.735119] env[61936]: DEBUG nova.network.neutron [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Updating instance_info_cache with network_info: [{"id": "853ddccf-d49b-471f-afc6-5cf28851a036", "address": "fa:16:3e:fd:9c:aa", "network": {"id": "5e8ea660-1601-4e5f-94a1-95119eb79202", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-655319130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c586842b2cf34c1e999dc110ba05261b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853ddccf-d4", "ovs_interfaceid": "853ddccf-d49b-471f-afc6-5cf28851a036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.739190] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00911701-e767-4a2f-8cfd-6723c5a32b8b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.749876] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3310f6-87b5-4d46-905d-48c16b71f312 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.782076] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d51277-d024-45ec-8fd4-960cf03d895b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.790233] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a5b330-b9d9-4319-84d2-93ee34bdd06f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.804360] env[61936]: DEBUG nova.compute.provider_tree [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.959644] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252884, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.008424] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252883, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.165971] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5283a374-40b6-ba70-dad1-b612642ee990, 'name': SearchDatastore_Task, 'duration_secs': 0.009057} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.166317] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.166640] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d/644ccb42-44da-43e8-8b3b-9bfd2ac19a5d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.166940] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a20042a1-c029-4bb2-bb20-33fb71aea188 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.173669] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 845.173669] env[61936]: value = "task-1252885" [ 845.173669] env[61936]: _type = "Task" [ 845.173669] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.181632] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252885, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.198405] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525fd798-fec4-4917-32ae-367f399f66d7, 'name': SearchDatastore_Task, 'duration_secs': 0.008553} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.198658] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.198990] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.199147] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.199303] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.199482] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.199731] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96be05ab-3474-4155-bc48-ff1bb2d68f41 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.207196] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.207333] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.208129] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fb51831-0b0e-4492-85a0-f0cd0200b5b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.212767] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 845.212767] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52899233-eeb9-3947-b0ad-9c3406cdc2e4" [ 845.212767] env[61936]: _type = "Task" [ 845.212767] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.219831] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52899233-eeb9-3947-b0ad-9c3406cdc2e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.237678] env[61936]: DEBUG oslo_concurrency.lockutils [req-2dd93802-4cda-4157-9eda-d837fbf99bdd req-9712d5e7-e957-4f75-9004-7951cc4d3434 service nova] Releasing lock "refresh_cache-fafc995e-4c93-4ca0-b078-24a1ae0ab427" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.310360] env[61936]: DEBUG nova.scheduler.client.report [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 845.375071] env[61936]: DEBUG nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 845.405479] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 845.405736] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.406128] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 845.406128] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.406277] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 845.406402] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 845.406608] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 845.406762] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 845.406938] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 845.407112] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 845.407286] env[61936]: DEBUG nova.virt.hardware [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 845.408245] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3209c01b-d934-4461-a915-757653ad5cb1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.416902] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4898f7-f396-44b2-a9b1-59b601851693 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.460261] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252884, 'name': Rename_Task, 'duration_secs': 0.891119} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.460479] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 845.460734] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db898e8a-8ea4-42f0-8c4f-bf11909b6f31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.470713] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 845.470713] env[61936]: value = "task-1252886" [ 845.470713] env[61936]: _type = "Task" [ 845.470713] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.479184] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.512735] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252883, 'name': ReconfigVM_Task, 'duration_secs': 1.236685} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.513224] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Reconfigured VM instance instance-00000038 to attach disk [datastore1] bdea3742-5789-4e97-b139-33cfee134843/bdea3742-5789-4e97-b139-33cfee134843.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.514036] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d03729f-6949-4690-92ce-553bdd3e5409 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.523629] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 845.523629] env[61936]: value = "task-1252887" [ 845.523629] env[61936]: _type = "Task" [ 845.523629] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.535537] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252887, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.684197] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252885, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487726} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.684471] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d/644ccb42-44da-43e8-8b3b-9bfd2ac19a5d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.684687] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.684936] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a0998959-41ee-4783-a587-0fe9e2391067 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.691353] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 845.691353] env[61936]: value = "task-1252888" [ 845.691353] env[61936]: _type = "Task" [ 845.691353] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.698923] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252888, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.723080] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52899233-eeb9-3947-b0ad-9c3406cdc2e4, 'name': SearchDatastore_Task, 'duration_secs': 0.012925} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.723964] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e33d6a5-e63d-43bf-a313-149051edbbd9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.729774] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 845.729774] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e17116-d04f-247e-5ea1-8b4a45908f62" [ 845.729774] env[61936]: _type = "Task" [ 845.729774] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.737724] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e17116-d04f-247e-5ea1-8b4a45908f62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.816209] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.816828] env[61936]: DEBUG nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 845.819912] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.670s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.820175] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.822453] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.979s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.824235] env[61936]: INFO nova.compute.claims [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.851484] env[61936]: INFO nova.scheduler.client.report [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Deleted allocations for instance 867a4bdc-8326-48b3-ba17-ca8336533c2d [ 845.942142] env[61936]: DEBUG nova.network.neutron [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Successfully updated port: ca6d9c8e-18c6-4a80-a792-baa44a0acc54 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.978719] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252886, 'name': PowerOnVM_Task, 'duration_secs': 0.492796} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.979014] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 845.979259] env[61936]: INFO nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Took 7.98 seconds to spawn the instance on the hypervisor. [ 845.979439] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 845.980221] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918084c9-d1d2-47bb-a740-51388cf85824 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.985429] env[61936]: DEBUG nova.compute.manager [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Received event network-vif-plugged-ca6d9c8e-18c6-4a80-a792-baa44a0acc54 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 845.985689] env[61936]: DEBUG oslo_concurrency.lockutils [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] Acquiring lock "e5e3e7af-8104-4f7a-ae18-81e03a932608-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.985834] env[61936]: DEBUG oslo_concurrency.lockutils [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.985998] env[61936]: DEBUG oslo_concurrency.lockutils [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.986191] env[61936]: DEBUG nova.compute.manager [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] No waiting events found dispatching network-vif-plugged-ca6d9c8e-18c6-4a80-a792-baa44a0acc54 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 845.986322] env[61936]: WARNING nova.compute.manager [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Received unexpected event network-vif-plugged-ca6d9c8e-18c6-4a80-a792-baa44a0acc54 for instance with vm_state building and task_state spawning. [ 845.986486] env[61936]: DEBUG nova.compute.manager [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Received event network-changed-ca6d9c8e-18c6-4a80-a792-baa44a0acc54 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 845.986638] env[61936]: DEBUG nova.compute.manager [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Refreshing instance network info cache due to event network-changed-ca6d9c8e-18c6-4a80-a792-baa44a0acc54. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 845.986871] env[61936]: DEBUG oslo_concurrency.lockutils [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] Acquiring lock "refresh_cache-e5e3e7af-8104-4f7a-ae18-81e03a932608" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.986940] env[61936]: DEBUG oslo_concurrency.lockutils [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] Acquired lock "refresh_cache-e5e3e7af-8104-4f7a-ae18-81e03a932608" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.987226] env[61936]: DEBUG nova.network.neutron [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Refreshing network info cache for port ca6d9c8e-18c6-4a80-a792-baa44a0acc54 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.034567] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252887, 'name': Rename_Task, 'duration_secs': 0.374438} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.034822] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.035068] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f8de9b8-6d0b-4cc1-9c4c-05637b9811ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.040790] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 846.040790] env[61936]: value = "task-1252889" [ 846.040790] env[61936]: _type = "Task" [ 846.040790] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.048313] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252889, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.202033] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252888, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069566} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.202033] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.202643] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d481de-b8f1-4651-b353-eb3e2bbbbd28 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.221967] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d/644ccb42-44da-43e8-8b3b-9bfd2ac19a5d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.222279] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2519935-6b71-4fbd-923b-dc34c422df3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.245215] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e17116-d04f-247e-5ea1-8b4a45908f62, 'name': SearchDatastore_Task, 'duration_secs': 0.008662} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.246574] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.246872] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fafc995e-4c93-4ca0-b078-24a1ae0ab427/fafc995e-4c93-4ca0-b078-24a1ae0ab427.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.247238] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 846.247238] env[61936]: value = "task-1252890" [ 846.247238] env[61936]: _type = "Task" [ 846.247238] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.247451] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9115ede2-ccfb-40c6-8f31-e7d5396b8a8e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.258152] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252890, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.259412] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 846.259412] env[61936]: value = "task-1252891" [ 846.259412] env[61936]: _type = "Task" [ 846.259412] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.267016] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.329358] env[61936]: DEBUG nova.compute.utils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.332856] env[61936]: DEBUG nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 846.332856] env[61936]: DEBUG nova.network.neutron [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.360562] env[61936]: DEBUG oslo_concurrency.lockutils [None req-817bcded-f13b-4b1a-868e-208ae29b65ea tempest-InstanceActionsV221TestJSON-1847332905 tempest-InstanceActionsV221TestJSON-1847332905-project-member] Lock "867a4bdc-8326-48b3-ba17-ca8336533c2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.863s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.384017] env[61936]: DEBUG nova.policy [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ab09da635d44a9992809c35449ba07d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c21a56ef57b4d78bb6b1e6343efc442', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.443701] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "refresh_cache-e5e3e7af-8104-4f7a-ae18-81e03a932608" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.503139] env[61936]: INFO nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Took 36.11 seconds to build instance. [ 846.544465] env[61936]: DEBUG nova.network.neutron [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.557607] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252889, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.639367] env[61936]: DEBUG nova.network.neutron [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.762102] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.770441] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252891, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.837336] env[61936]: DEBUG nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 847.000377] env[61936]: DEBUG nova.network.neutron [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Successfully created port: dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.005266] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.371s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.061024] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252889, 'name': PowerOnVM_Task, 'duration_secs': 0.988949} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.061523] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.061748] env[61936]: INFO nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Took 11.55 seconds to spawn the instance on the hypervisor. [ 847.061956] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 847.062896] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca86a3d6-1a3a-4aa8-a83f-c22ed2ea0165 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.141567] env[61936]: DEBUG oslo_concurrency.lockutils [req-1573ed0a-8bf0-49bf-b46f-4ce2bb14182e req-7cb09b56-d8f1-42a1-af11-11df8bd49b82 service nova] Releasing lock "refresh_cache-e5e3e7af-8104-4f7a-ae18-81e03a932608" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.141976] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquired lock "refresh_cache-e5e3e7af-8104-4f7a-ae18-81e03a932608" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.142153] env[61936]: DEBUG nova.network.neutron [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.244259] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775093f0-3418-418d-ae3a-4c4e054be9df {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.255067] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4633293-5412-4882-95f8-e5bae2cf5973 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.290306] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252890, 'name': ReconfigVM_Task, 'duration_secs': 1.017044} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.291454] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c842521d-5388-46b1-b5f2-6fc5dffb1399 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.294878] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d/644ccb42-44da-43e8-8b3b-9bfd2ac19a5d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.298495] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1ae6230-34be-4697-bbf5-97113d42e59f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.300074] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.964319} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.300649] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fafc995e-4c93-4ca0-b078-24a1ae0ab427/fafc995e-4c93-4ca0-b078-24a1ae0ab427.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.300855] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.301372] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4af66857-e264-4437-9390-d855b4ba1b99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.308032] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f99a8a-fa2f-45b0-8eec-b2daf437e0b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.312172] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 847.312172] env[61936]: value = "task-1252892" [ 847.312172] env[61936]: _type = "Task" [ 847.312172] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.313625] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 847.313625] env[61936]: value = "task-1252893" [ 847.313625] env[61936]: _type = "Task" [ 847.313625] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.331693] env[61936]: DEBUG nova.compute.provider_tree [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.337781] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252892, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.340656] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252893, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.512946] env[61936]: DEBUG nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 847.585090] env[61936]: INFO nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Took 39.13 seconds to build instance. [ 847.689534] env[61936]: DEBUG nova.network.neutron [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.825364] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252892, 'name': Rename_Task, 'duration_secs': 0.196217} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.828200] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.828457] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252893, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067144} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.828648] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-054d546e-ea13-40b0-ba83-3509fe1693f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.830053] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.830794] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da21686-2939-4809-a8b4-d4d3abe0dbfd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.836081] env[61936]: DEBUG nova.scheduler.client.report [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 847.852152] env[61936]: DEBUG nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 847.862496] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] fafc995e-4c93-4ca0-b078-24a1ae0ab427/fafc995e-4c93-4ca0-b078-24a1ae0ab427.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.863563] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 847.863563] env[61936]: value = "task-1252894" [ 847.863563] env[61936]: _type = "Task" [ 847.863563] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.867815] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f327ad4-2767-465c-ac1a-98300676fcff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.900844] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252894, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.903258] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 847.903784] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.904141] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 847.904475] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.904755] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 847.905057] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 847.905415] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 847.905732] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 847.906048] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 847.906360] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 847.906681] env[61936]: DEBUG nova.virt.hardware [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 847.908657] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a868f576-ad7c-40f6-926c-cca1731df10d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.911752] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 847.911752] env[61936]: value = "task-1252895" [ 847.911752] env[61936]: _type = "Task" [ 847.911752] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.920904] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43248d7-820e-495b-8f29-296a02b0df07 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.927691] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.953780] env[61936]: DEBUG nova.network.neutron [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Updating instance_info_cache with network_info: [{"id": "ca6d9c8e-18c6-4a80-a792-baa44a0acc54", "address": "fa:16:3e:14:11:5f", "network": {"id": "389de814-9758-4590-b955-a9e9d7738e97", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1179729655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a643413e3d7748cbac250b824fcbba39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca6d9c8e-18", "ovs_interfaceid": "ca6d9c8e-18c6-4a80-a792-baa44a0acc54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.036659] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.088792] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "bdea3742-5789-4e97-b139-33cfee134843" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.489s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.364378] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.364895] env[61936]: DEBUG nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 848.369793] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.979s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.373020] env[61936]: INFO nova.compute.claims [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.396391] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252894, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.423426] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.456518] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Releasing lock "refresh_cache-e5e3e7af-8104-4f7a-ae18-81e03a932608" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.456855] env[61936]: DEBUG nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Instance network_info: |[{"id": "ca6d9c8e-18c6-4a80-a792-baa44a0acc54", "address": "fa:16:3e:14:11:5f", "network": {"id": "389de814-9758-4590-b955-a9e9d7738e97", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1179729655-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a643413e3d7748cbac250b824fcbba39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca6d9c8e-18", "ovs_interfaceid": "ca6d9c8e-18c6-4a80-a792-baa44a0acc54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 848.457327] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:11:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca6d9c8e-18c6-4a80-a792-baa44a0acc54', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.464975] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Creating folder: Project (a643413e3d7748cbac250b824fcbba39). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.465330] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a7c49d5-e2ec-4bd6-aa9c-e14b3b984f54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.477029] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Created folder: Project (a643413e3d7748cbac250b824fcbba39) in parent group-v269874. [ 848.477267] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Creating folder: Instances. Parent ref: group-v269927. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.477524] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89ebcd99-7137-4522-b505-93c788225ba3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.491936] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Created folder: Instances in parent group-v269927. [ 848.492310] env[61936]: DEBUG oslo.service.loopingcall [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.492678] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.492944] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5ffc8a0-bea9-4391-9e28-ccd666431462 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.514431] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.514431] env[61936]: value = "task-1252898" [ 848.514431] env[61936]: _type = "Task" [ 848.514431] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.530286] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252898, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.593563] env[61936]: DEBUG nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 848.878021] env[61936]: DEBUG nova.compute.utils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.886022] env[61936]: DEBUG nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 848.886022] env[61936]: DEBUG nova.network.neutron [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.901196] env[61936]: DEBUG oslo_vmware.api [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252894, 'name': PowerOnVM_Task, 'duration_secs': 0.69211} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.902762] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.903159] env[61936]: INFO nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Took 5.91 seconds to spawn the instance on the hypervisor. [ 848.905891] env[61936]: DEBUG nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 848.905891] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a46bb4-ed6e-40a5-be6d-a2b892356983 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.929926] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252895, 'name': ReconfigVM_Task, 'duration_secs': 0.598397} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.931875] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Reconfigured VM instance instance-0000003a to attach disk [datastore2] fafc995e-4c93-4ca0-b078-24a1ae0ab427/fafc995e-4c93-4ca0-b078-24a1ae0ab427.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.931875] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ee404a9-8bcb-437b-90fb-b0e2bb422326 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.939069] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 848.939069] env[61936]: value = "task-1252899" [ 848.939069] env[61936]: _type = "Task" [ 848.939069] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.949837] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252899, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.974467] env[61936]: DEBUG nova.policy [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '288fd1d0d8e341999481fc62fe14195e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f1b64c32564b9cbb68e3799b456641', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 849.013468] env[61936]: DEBUG nova.compute.manager [req-12e1a7a4-b6a6-42da-a848-c9cad473d266 req-700129b2-ec88-4a0e-8ca4-0fe3c542e712 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Received event network-vif-plugged-dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 849.013694] env[61936]: DEBUG oslo_concurrency.lockutils [req-12e1a7a4-b6a6-42da-a848-c9cad473d266 req-700129b2-ec88-4a0e-8ca4-0fe3c542e712 service nova] Acquiring lock "f2844173-4114-4158-99ad-ac651efbb607-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.013896] env[61936]: DEBUG oslo_concurrency.lockutils [req-12e1a7a4-b6a6-42da-a848-c9cad473d266 req-700129b2-ec88-4a0e-8ca4-0fe3c542e712 service nova] Lock "f2844173-4114-4158-99ad-ac651efbb607-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.014074] env[61936]: DEBUG oslo_concurrency.lockutils [req-12e1a7a4-b6a6-42da-a848-c9cad473d266 req-700129b2-ec88-4a0e-8ca4-0fe3c542e712 service nova] Lock "f2844173-4114-4158-99ad-ac651efbb607-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.014772] env[61936]: DEBUG nova.compute.manager [req-12e1a7a4-b6a6-42da-a848-c9cad473d266 req-700129b2-ec88-4a0e-8ca4-0fe3c542e712 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] No waiting events found dispatching network-vif-plugged-dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 849.014772] env[61936]: WARNING nova.compute.manager [req-12e1a7a4-b6a6-42da-a848-c9cad473d266 req-700129b2-ec88-4a0e-8ca4-0fe3c542e712 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Received unexpected event network-vif-plugged-dd14ed80-5888-44af-897b-0ed4dd854e0d for instance with vm_state building and task_state spawning. [ 849.025437] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252898, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.115321] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.307479] env[61936]: DEBUG nova.network.neutron [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Successfully created port: 48edeab7-9688-4106-9b4a-9e53a69aa280 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.383544] env[61936]: DEBUG nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 849.434891] env[61936]: INFO nova.compute.manager [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Took 36.23 seconds to build instance. [ 849.454777] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252899, 'name': Rename_Task, 'duration_secs': 0.293487} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.459610] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.462478] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bfc61db-83e5-4bdb-a02e-bf10f3c0bfb5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.472035] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 849.472035] env[61936]: value = "task-1252900" [ 849.472035] env[61936]: _type = "Task" [ 849.472035] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.486802] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252900, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.493693] env[61936]: DEBUG nova.network.neutron [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Successfully updated port: dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.539355] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252898, 'name': CreateVM_Task, 'duration_secs': 0.546626} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.539677] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.541372] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.541372] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.541372] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.541372] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f38e212-d83c-45d2-863e-5c2f9ae780ff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.547705] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 849.547705] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b51ca-1c33-3ab6-3813-02d708f09c81" [ 849.547705] env[61936]: _type = "Task" [ 849.547705] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.557514] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b51ca-1c33-3ab6-3813-02d708f09c81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.658879] env[61936]: DEBUG nova.compute.manager [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Received event network-changed-dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 849.659114] env[61936]: DEBUG nova.compute.manager [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Refreshing instance network info cache due to event network-changed-dd14ed80-5888-44af-897b-0ed4dd854e0d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 849.659345] env[61936]: DEBUG oslo_concurrency.lockutils [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] Acquiring lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.659493] env[61936]: DEBUG oslo_concurrency.lockutils [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] Acquired lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.659651] env[61936]: DEBUG nova.network.neutron [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Refreshing network info cache for port dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.793398] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a409b606-15f4-498f-9330-f782bcd242e9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.801808] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecbe819-956d-4a2f-a219-efbb8c51947d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.834046] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3435526f-f9b7-4614-84a2-30d60733cdb0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.841737] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb37e210-4dcd-4718-85ec-30523608a530 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.855589] env[61936]: DEBUG nova.compute.provider_tree [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.936881] env[61936]: DEBUG oslo_concurrency.lockutils [None req-99a26e51-fcfe-4660-9cf8-902c5f69bb00 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.380s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.985511] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252900, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.996311] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.058133] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b51ca-1c33-3ab6-3813-02d708f09c81, 'name': SearchDatastore_Task, 'duration_secs': 0.011829} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.058465] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.058725] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.058958] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.059119] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.059294] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.059556] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b33e9ce6-7d9c-498c-8b45-d25f80f4724a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.073401] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.073498] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.074331] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f85c4127-d588-41a1-abf1-c4c2046c11fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.078872] env[61936]: DEBUG nova.compute.manager [None req-80463b6b-8152-452f-a4eb-7828dd34809f tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 850.081946] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab71a432-b447-4074-98c0-51893140cc90 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.086219] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 850.086219] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5224b082-2c26-f7e0-83c5-b6a921c9bda4" [ 850.086219] env[61936]: _type = "Task" [ 850.086219] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.102680] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5224b082-2c26-f7e0-83c5-b6a921c9bda4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.197469] env[61936]: DEBUG nova.network.neutron [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.214790] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.214790] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.214986] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.215890] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.215890] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.217798] env[61936]: INFO nova.compute.manager [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Terminating instance [ 850.285362] env[61936]: DEBUG nova.network.neutron [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.359323] env[61936]: DEBUG nova.scheduler.client.report [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 850.394783] env[61936]: DEBUG nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 850.422014] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 850.422266] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.422422] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 850.422598] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.422739] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 850.422878] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 850.423091] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 850.423251] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 850.423414] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 850.423578] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 850.423747] env[61936]: DEBUG nova.virt.hardware [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 850.424692] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07564620-5871-4dae-a246-5ddabbfe4630 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.435050] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ebeeb9-de03-4732-8b67-62bfcf8364b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.439978] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 850.500049] env[61936]: DEBUG oslo_vmware.api [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252900, 'name': PowerOnVM_Task, 'duration_secs': 0.718501} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.500049] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.500049] env[61936]: INFO nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Took 10.03 seconds to spawn the instance on the hypervisor. [ 850.500049] env[61936]: DEBUG nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 850.500049] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d7aeea-e6bc-451c-9fe2-1d88f595c25f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.599021] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5224b082-2c26-f7e0-83c5-b6a921c9bda4, 'name': SearchDatastore_Task, 'duration_secs': 0.025303} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.599021] env[61936]: INFO nova.compute.manager [None req-80463b6b-8152-452f-a4eb-7828dd34809f tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] instance snapshotting [ 850.599021] env[61936]: DEBUG nova.objects.instance [None req-80463b6b-8152-452f-a4eb-7828dd34809f tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lazy-loading 'flavor' on Instance uuid 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.600113] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83f157c5-fd90-46e2-b7ec-dff8a8ca8dd7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.605667] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 850.605667] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5287be9f-8ecb-d698-a885-4c198bc577da" [ 850.605667] env[61936]: _type = "Task" [ 850.605667] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.615041] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5287be9f-8ecb-d698-a885-4c198bc577da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.723848] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "refresh_cache-644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.723848] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquired lock "refresh_cache-644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.723848] env[61936]: DEBUG nova.network.neutron [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.777858] env[61936]: DEBUG nova.compute.manager [req-920e3c9b-de38-4ff4-be78-23b8e6787da9 req-46d1aa92-9834-4f37-ad87-ab50c3d1743b service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Received event network-vif-plugged-48edeab7-9688-4106-9b4a-9e53a69aa280 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 850.779277] env[61936]: DEBUG oslo_concurrency.lockutils [req-920e3c9b-de38-4ff4-be78-23b8e6787da9 req-46d1aa92-9834-4f37-ad87-ab50c3d1743b service nova] Acquiring lock "e800f5e9-deb0-4592-aded-3768d238e311-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.779277] env[61936]: DEBUG oslo_concurrency.lockutils [req-920e3c9b-de38-4ff4-be78-23b8e6787da9 req-46d1aa92-9834-4f37-ad87-ab50c3d1743b service nova] Lock "e800f5e9-deb0-4592-aded-3768d238e311-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.779277] env[61936]: DEBUG oslo_concurrency.lockutils [req-920e3c9b-de38-4ff4-be78-23b8e6787da9 req-46d1aa92-9834-4f37-ad87-ab50c3d1743b service nova] Lock "e800f5e9-deb0-4592-aded-3768d238e311-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.779895] env[61936]: DEBUG nova.compute.manager [req-920e3c9b-de38-4ff4-be78-23b8e6787da9 req-46d1aa92-9834-4f37-ad87-ab50c3d1743b service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] No waiting events found dispatching network-vif-plugged-48edeab7-9688-4106-9b4a-9e53a69aa280 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 850.780257] env[61936]: WARNING nova.compute.manager [req-920e3c9b-de38-4ff4-be78-23b8e6787da9 req-46d1aa92-9834-4f37-ad87-ab50c3d1743b service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Received unexpected event network-vif-plugged-48edeab7-9688-4106-9b4a-9e53a69aa280 for instance with vm_state building and task_state spawning. [ 850.788683] env[61936]: DEBUG oslo_concurrency.lockutils [req-443e00db-51fd-412f-8b87-de1baecc9707 req-59abe82b-1ad5-4ecc-8afd-87afc91a6594 service nova] Releasing lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.788683] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.788859] env[61936]: DEBUG nova.network.neutron [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.869165] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.869165] env[61936]: DEBUG nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 850.877020] env[61936]: DEBUG nova.network.neutron [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Successfully updated port: 48edeab7-9688-4106-9b4a-9e53a69aa280 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.877020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.411s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.877020] env[61936]: DEBUG nova.objects.instance [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lazy-loading 'resources' on Instance uuid 50cb3f79-39b2-46ab-a35c-fdad3bb266cd {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.974702] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.021046] env[61936]: INFO nova.compute.manager [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Took 40.34 seconds to build instance. [ 851.109367] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb457ca-ac3d-451b-9bf6-dfccda03e18f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.128635] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5287be9f-8ecb-d698-a885-4c198bc577da, 'name': SearchDatastore_Task, 'duration_secs': 0.024575} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.146385] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.146692] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] e5e3e7af-8104-4f7a-ae18-81e03a932608/e5e3e7af-8104-4f7a-ae18-81e03a932608.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.147162] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47904a98-5871-4886-aa2e-e7befd565b3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.149918] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711d4a22-bef6-49f0-949c-8bb4c129bd9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.165192] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 851.165192] env[61936]: value = "task-1252901" [ 851.165192] env[61936]: _type = "Task" [ 851.165192] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.174556] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252901, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.243699] env[61936]: DEBUG nova.network.neutron [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.300113] env[61936]: DEBUG nova.network.neutron [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.356227] env[61936]: DEBUG nova.network.neutron [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.381054] env[61936]: DEBUG nova.compute.utils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.383235] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-e800f5e9-deb0-4592-aded-3768d238e311" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.383235] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-e800f5e9-deb0-4592-aded-3768d238e311" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.383235] env[61936]: DEBUG nova.network.neutron [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.389192] env[61936]: DEBUG nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 851.389373] env[61936]: DEBUG nova.network.neutron [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.447541] env[61936]: DEBUG nova.policy [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.523511] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c9b8c9c7-f281-4fc5-ad1f-e154a50f173b tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.855s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.664106] env[61936]: DEBUG nova.compute.manager [None req-80463b6b-8152-452f-a4eb-7828dd34809f tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Instance disappeared during snapshot {{(pid=61936) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 851.678207] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252901, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.793487] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee98c854-dde9-43f9-b3dc-653aca8fdfae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.804412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Releasing lock "refresh_cache-644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.804814] env[61936]: DEBUG nova.compute.manager [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 851.805009] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 851.805833] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa006e0b-5209-4ec8-be22-8d9135052e3c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.810214] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0282d9-1b0b-4798-b94d-33547e2aa2b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.819058] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 851.820388] env[61936]: DEBUG nova.network.neutron [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Updating instance_info_cache with network_info: [{"id": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "address": "fa:16:3e:87:2f:dd", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd14ed80-58", "ovs_interfaceid": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.850358] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b72df143-c5c7-4861-8f3d-040aa0c79f80 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.858482] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Releasing lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.858998] env[61936]: DEBUG nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Instance network_info: |[{"id": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "address": "fa:16:3e:87:2f:dd", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd14ed80-58", "ovs_interfaceid": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 851.859656] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f775243d-729c-4128-913f-b33aec169604 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.862650] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:2f:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b033f4d-2e92-4702-add6-410a29d3f251', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd14ed80-5888-44af-897b-0ed4dd854e0d', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.870369] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Creating folder: Project (4c21a56ef57b4d78bb6b1e6343efc442). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.871626] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f0446d1-de0d-40b3-a3ae-f71f3b398bed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.876281] env[61936]: DEBUG oslo_vmware.api [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 851.876281] env[61936]: value = "task-1252902" [ 851.876281] env[61936]: _type = "Task" [ 851.876281] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.883164] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1536e7-114c-4c3e-8ddb-0ec53cc1e003 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.890737] env[61936]: DEBUG nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 851.907077] env[61936]: DEBUG oslo_vmware.api [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.907905] env[61936]: DEBUG nova.compute.provider_tree [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.912310] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Created folder: Project (4c21a56ef57b4d78bb6b1e6343efc442) in parent group-v269874. [ 851.912310] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Creating folder: Instances. Parent ref: group-v269930. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.912310] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-270c063b-5c86-42b5-b7db-563315664dcc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.921232] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Created folder: Instances in parent group-v269930. [ 851.921634] env[61936]: DEBUG oslo.service.loopingcall [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.922456] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2844173-4114-4158-99ad-ac651efbb607] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.923444] env[61936]: DEBUG nova.compute.manager [None req-80463b6b-8152-452f-a4eb-7828dd34809f tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Found 0 images (rotation: 2) {{(pid=61936) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4941}} [ 851.924379] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a914472-93ca-44ba-aa10-a58eaecb0c00 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.945092] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.945092] env[61936]: value = "task-1252905" [ 851.945092] env[61936]: _type = "Task" [ 851.945092] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.953197] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252905, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.965642] env[61936]: DEBUG nova.network.neutron [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.995028] env[61936]: DEBUG nova.network.neutron [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Successfully created port: 12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.033011] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 852.177996] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252901, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600613} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.178429] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] e5e3e7af-8104-4f7a-ae18-81e03a932608/e5e3e7af-8104-4f7a-ae18-81e03a932608.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.178565] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.180872] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c89f223c-a133-4775-9d60-44c1d8f6ace8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.188078] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 852.188078] env[61936]: value = "task-1252906" [ 852.188078] env[61936]: _type = "Task" [ 852.188078] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.193080] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252906, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.244113] env[61936]: DEBUG nova.network.neutron [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Updating instance_info_cache with network_info: [{"id": "48edeab7-9688-4106-9b4a-9e53a69aa280", "address": "fa:16:3e:a7:2f:d7", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48edeab7-96", "ovs_interfaceid": "48edeab7-9688-4106-9b4a-9e53a69aa280", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.391373] env[61936]: DEBUG oslo_vmware.api [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252902, 'name': PowerOffVM_Task, 'duration_secs': 0.250197} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.394453] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.394453] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 852.394453] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac14690c-ca7b-4bee-af9e-e06394424a90 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.415080] env[61936]: DEBUG nova.scheduler.client.report [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 852.421389] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 852.421605] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 852.421779] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Deleting the datastore file [datastore2] 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.422052] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5008ad14-6d31-4016-8797-961ff7ebe696 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.429806] env[61936]: DEBUG oslo_vmware.api [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for the task: (returnval){ [ 852.429806] env[61936]: value = "task-1252908" [ 852.429806] env[61936]: _type = "Task" [ 852.429806] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.444917] env[61936]: DEBUG oslo_vmware.api [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252908, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.457346] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252905, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.564183] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.696918] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252906, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070025} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.697216] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.698276] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a47dda-df39-4497-8afe-05e4bec7e8e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.723057] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] e5e3e7af-8104-4f7a-ae18-81e03a932608/e5e3e7af-8104-4f7a-ae18-81e03a932608.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.724769] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13d438e6-35c6-4ff8-a4d1-c7c7b7b9f517 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.747200] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-e800f5e9-deb0-4592-aded-3768d238e311" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.747575] env[61936]: DEBUG nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Instance network_info: |[{"id": "48edeab7-9688-4106-9b4a-9e53a69aa280", "address": "fa:16:3e:a7:2f:d7", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48edeab7-96", "ovs_interfaceid": "48edeab7-9688-4106-9b4a-9e53a69aa280", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 852.747993] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:2f:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b94712a6-b777-47dd-bc06-f9acfce2d936', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48edeab7-9688-4106-9b4a-9e53a69aa280', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.757365] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Creating folder: Project (d4f1b64c32564b9cbb68e3799b456641). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.761247] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7de492d-7706-4a76-9f41-5855987deb0b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.763446] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 852.763446] env[61936]: value = "task-1252909" [ 852.763446] env[61936]: _type = "Task" [ 852.763446] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.780103] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252909, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.780103] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Created folder: Project (d4f1b64c32564b9cbb68e3799b456641) in parent group-v269874. [ 852.780103] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Creating folder: Instances. Parent ref: group-v269933. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.780103] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03c2e9fd-93a7-4599-aefb-66dd8d5e51a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.788139] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Created folder: Instances in parent group-v269933. [ 852.788396] env[61936]: DEBUG oslo.service.loopingcall [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.788616] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.788871] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0041d799-1c1d-4044-8c91-60028131d3a1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.809428] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.809428] env[61936]: value = "task-1252912" [ 852.809428] env[61936]: _type = "Task" [ 852.809428] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.817467] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252912, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.920988] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.044s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.924043] env[61936]: DEBUG nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 852.926490] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.968s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.928014] env[61936]: INFO nova.compute.claims [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.945988] env[61936]: DEBUG oslo_vmware.api [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Task: {'id': task-1252908, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.28579} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.947815] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.948202] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 852.948287] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 852.948444] env[61936]: INFO nova.compute.manager [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 852.948674] env[61936]: DEBUG oslo.service.loopingcall [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.954021] env[61936]: DEBUG nova.compute.manager [-] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 852.954073] env[61936]: DEBUG nova.network.neutron [-] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 852.957032] env[61936]: INFO nova.scheduler.client.report [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Deleted allocations for instance 50cb3f79-39b2-46ab-a35c-fdad3bb266cd [ 852.967794] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 852.968080] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.968284] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 852.968477] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.968593] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 852.968738] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 852.968943] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 852.969144] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 852.969338] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 852.969504] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 852.969686] env[61936]: DEBUG nova.virt.hardware [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 852.976951] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ac2138-0836-43b8-901d-53bf6374de88 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.980232] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252905, 'name': CreateVM_Task, 'duration_secs': 0.702908} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.980667] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2844173-4114-4158-99ad-ac651efbb607] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.981792] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.982021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.982274] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.982536] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-679c2e89-fafe-43fa-a7cd-6f5bf666d7e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.986584] env[61936]: DEBUG nova.network.neutron [-] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.995100] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ca67cf-86fd-4d9d-b3fb-5818bf9833cc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.006099] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 853.006099] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52730336-34ee-091a-140e-48a2a4cecddf" [ 853.006099] env[61936]: _type = "Task" [ 853.006099] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.022775] env[61936]: DEBUG nova.compute.manager [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Received event network-changed-48edeab7-9688-4106-9b4a-9e53a69aa280 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 853.023037] env[61936]: DEBUG nova.compute.manager [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Refreshing instance network info cache due to event network-changed-48edeab7-9688-4106-9b4a-9e53a69aa280. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 853.024255] env[61936]: DEBUG oslo_concurrency.lockutils [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] Acquiring lock "refresh_cache-e800f5e9-deb0-4592-aded-3768d238e311" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.024255] env[61936]: DEBUG oslo_concurrency.lockutils [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] Acquired lock "refresh_cache-e800f5e9-deb0-4592-aded-3768d238e311" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.024255] env[61936]: DEBUG nova.network.neutron [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Refreshing network info cache for port 48edeab7-9688-4106-9b4a-9e53a69aa280 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.029487] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52730336-34ee-091a-140e-48a2a4cecddf, 'name': SearchDatastore_Task, 'duration_secs': 0.019899} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.029960] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.030229] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.030452] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.030588] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.030942] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.031431] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-977ecd87-9b94-4889-88d4-f266df7f24b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.041095] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.041295] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.042072] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34e6dd8c-f699-4862-9da5-65af4b937376 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.047812] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 853.047812] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f5ad9f-93d9-390d-97f8-6a9eb004884d" [ 853.047812] env[61936]: _type = "Task" [ 853.047812] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.056273] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f5ad9f-93d9-390d-97f8-6a9eb004884d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.244884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "bdea3742-5789-4e97-b139-33cfee134843" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.245102] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "bdea3742-5789-4e97-b139-33cfee134843" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.245310] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "bdea3742-5789-4e97-b139-33cfee134843-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.245484] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "bdea3742-5789-4e97-b139-33cfee134843-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.245642] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "bdea3742-5789-4e97-b139-33cfee134843-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.250717] env[61936]: INFO nova.compute.manager [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Terminating instance [ 853.274349] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "301b0504-5cf9-44e0-bd3e-342f8a89278b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.274601] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.274796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "301b0504-5cf9-44e0-bd3e-342f8a89278b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.274970] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.275158] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.276788] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252909, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.277291] env[61936]: INFO nova.compute.manager [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Terminating instance [ 853.318803] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252912, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.484141] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d145163b-5258-4db1-a6eb-a1be71c2c533 tempest-ListImageFiltersTestJSON-1521853990 tempest-ListImageFiltersTestJSON-1521853990-project-member] Lock "50cb3f79-39b2-46ab-a35c-fdad3bb266cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.127s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.493700] env[61936]: DEBUG nova.network.neutron [-] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.560412] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f5ad9f-93d9-390d-97f8-6a9eb004884d, 'name': SearchDatastore_Task, 'duration_secs': 0.040652} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.561215] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acb96e40-101e-4079-b894-70d046327c6e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.570661] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 853.570661] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52033ded-a73c-52d1-aa03-e9214a66b984" [ 853.570661] env[61936]: _type = "Task" [ 853.570661] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.582645] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52033ded-a73c-52d1-aa03-e9214a66b984, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.754950] env[61936]: DEBUG nova.compute.manager [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 853.755210] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.756138] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d5038a-a61f-420b-b5a5-46aa693a85ee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.763931] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.764198] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70388bd4-89fb-498a-8ce6-8d105a5544c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.781167] env[61936]: DEBUG nova.compute.manager [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 853.781383] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.781651] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252909, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.785429] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4e517d-8813-4eb9-84c4-d83272b9c871 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.788377] env[61936]: DEBUG oslo_vmware.api [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 853.788377] env[61936]: value = "task-1252913" [ 853.788377] env[61936]: _type = "Task" [ 853.788377] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.795851] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.796616] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22b8dc72-f55d-4726-ad62-fd6c852ad586 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.801555] env[61936]: DEBUG oslo_vmware.api [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.806445] env[61936]: DEBUG oslo_vmware.api [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 853.806445] env[61936]: value = "task-1252914" [ 853.806445] env[61936]: _type = "Task" [ 853.806445] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.819952] env[61936]: DEBUG oslo_vmware.api [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.825709] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252912, 'name': CreateVM_Task, 'duration_secs': 0.565907} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.825968] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.826654] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.826836] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.827236] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.827694] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa118752-ccf7-448f-9eb6-b730ced1b175 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.832817] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 853.832817] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520246b8-cdde-cc72-a0c6-908f0b2fc576" [ 853.832817] env[61936]: _type = "Task" [ 853.832817] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.845026] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520246b8-cdde-cc72-a0c6-908f0b2fc576, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.853336] env[61936]: DEBUG nova.network.neutron [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Updated VIF entry in instance network info cache for port 48edeab7-9688-4106-9b4a-9e53a69aa280. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.853952] env[61936]: DEBUG nova.network.neutron [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Updating instance_info_cache with network_info: [{"id": "48edeab7-9688-4106-9b4a-9e53a69aa280", "address": "fa:16:3e:a7:2f:d7", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48edeab7-96", "ovs_interfaceid": "48edeab7-9688-4106-9b4a-9e53a69aa280", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.996876] env[61936]: INFO nova.compute.manager [-] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Took 1.04 seconds to deallocate network for instance. [ 854.084982] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52033ded-a73c-52d1-aa03-e9214a66b984, 'name': SearchDatastore_Task, 'duration_secs': 0.017832} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.085995] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.085995] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f2844173-4114-4158-99ad-ac651efbb607/f2844173-4114-4158-99ad-ac651efbb607.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.086243] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbc15365-22cd-4798-9756-19250bc9bb36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.092395] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 854.092395] env[61936]: value = "task-1252915" [ 854.092395] env[61936]: _type = "Task" [ 854.092395] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.106315] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252915, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.130698] env[61936]: DEBUG nova.network.neutron [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Successfully updated port: 12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.283357] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252909, 'name': ReconfigVM_Task, 'duration_secs': 1.335989} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.286250] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Reconfigured VM instance instance-0000003c to attach disk [datastore2] e5e3e7af-8104-4f7a-ae18-81e03a932608/e5e3e7af-8104-4f7a-ae18-81e03a932608.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.287156] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ad63eae-4c32-42f0-aa97-9da3677459ff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.297812] env[61936]: DEBUG oslo_vmware.api [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252913, 'name': PowerOffVM_Task, 'duration_secs': 0.264313} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.301979] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.302268] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.302614] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 854.302614] env[61936]: value = "task-1252916" [ 854.302614] env[61936]: _type = "Task" [ 854.302614] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.303898] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6220eadc-afdb-4882-b29b-938a53bde2de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.317210] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252916, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.324652] env[61936]: DEBUG oslo_vmware.api [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252914, 'name': PowerOffVM_Task, 'duration_secs': 0.213893} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.325499] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.325704] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.326956] env[61936]: DEBUG oslo_concurrency.lockutils [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.327185] env[61936]: DEBUG oslo_concurrency.lockutils [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.328468] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b4ec148-ec84-4c61-81b2-0e68e360131d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.349328] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520246b8-cdde-cc72-a0c6-908f0b2fc576, 'name': SearchDatastore_Task, 'duration_secs': 0.011243} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.349631] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.349862] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.350106] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.350503] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.350503] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.350761] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5033754d-7101-4b22-989a-48347f4c24a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.356355] env[61936]: DEBUG oslo_concurrency.lockutils [req-9feadd7c-8317-4ffa-967d-9824de2c76b3 req-c2a4c3be-93bf-4958-bf6d-12ab231dfcb9 service nova] Releasing lock "refresh_cache-e800f5e9-deb0-4592-aded-3768d238e311" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.379413] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.379668] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.381458] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b68ff106-956a-4140-a883-80df397b776c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.384625] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496c2136-8010-4421-b801-cc4722c4435b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.395732] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 854.395732] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52afb18e-21d3-84b8-b81b-98a79abb894a" [ 854.395732] env[61936]: _type = "Task" [ 854.395732] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.403501] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.403821] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.403997] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleting the datastore file [datastore1] bdea3742-5789-4e97-b139-33cfee134843 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.405221] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f2782c3-6813-415f-bd8b-794c4248b883 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.412983] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52afb18e-21d3-84b8-b81b-98a79abb894a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.415413] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a8b3aa-dcf1-4276-bcdb-d28b042831a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.418900] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.419118] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.419289] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Deleting the datastore file [datastore2] 301b0504-5cf9-44e0-bd3e-342f8a89278b {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.420741] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-445ae67e-b83c-4d59-909c-0d4f4b519262 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.422679] env[61936]: DEBUG oslo_vmware.api [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 854.422679] env[61936]: value = "task-1252919" [ 854.422679] env[61936]: _type = "Task" [ 854.422679] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.455612] env[61936]: DEBUG oslo_vmware.api [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for the task: (returnval){ [ 854.455612] env[61936]: value = "task-1252920" [ 854.455612] env[61936]: _type = "Task" [ 854.455612] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.457043] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a56e07-2dea-44bd-8190-710ad0565d11 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.467387] env[61936]: DEBUG oslo_vmware.api [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252919, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.477418] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283f9854-3803-492d-b897-b99421cd7450 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.482712] env[61936]: DEBUG oslo_vmware.api [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.494618] env[61936]: DEBUG nova.compute.provider_tree [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.512045] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.602772] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252915, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489072} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.603128] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f2844173-4114-4158-99ad-ac651efbb607/f2844173-4114-4158-99ad-ac651efbb607.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.603398] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.603659] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c211d7be-c0a4-42b0-84de-7be85e5bdaae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.611709] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 854.611709] env[61936]: value = "task-1252921" [ 854.611709] env[61936]: _type = "Task" [ 854.611709] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.620089] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252921, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.636637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.636637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.636637] env[61936]: DEBUG nova.network.neutron [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.817036] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252916, 'name': Rename_Task, 'duration_secs': 0.201717} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.817316] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.817709] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fca51b2-a126-4176-a4fc-88d5faccdb57 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.824945] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 854.824945] env[61936]: value = "task-1252922" [ 854.824945] env[61936]: _type = "Task" [ 854.824945] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.833404] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.836390] env[61936]: DEBUG nova.compute.utils [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.907878] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52afb18e-21d3-84b8-b81b-98a79abb894a, 'name': SearchDatastore_Task, 'duration_secs': 0.048626} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.908759] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-686ef4b3-eb36-4d7b-a916-2b432e8e0920 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.914093] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 854.914093] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b8092b-6105-c882-90af-696113520bd6" [ 854.914093] env[61936]: _type = "Task" [ 854.914093] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.924070] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b8092b-6105-c882-90af-696113520bd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.936416] env[61936]: DEBUG oslo_vmware.api [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1252919, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.271454} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.936725] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.936918] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 854.937108] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 854.937286] env[61936]: INFO nova.compute.manager [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: bdea3742-5789-4e97-b139-33cfee134843] Took 1.18 seconds to destroy the instance on the hypervisor. [ 854.937525] env[61936]: DEBUG oslo.service.loopingcall [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.937796] env[61936]: DEBUG nova.compute.manager [-] [instance: bdea3742-5789-4e97-b139-33cfee134843] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 854.937896] env[61936]: DEBUG nova.network.neutron [-] [instance: bdea3742-5789-4e97-b139-33cfee134843] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.972467] env[61936]: DEBUG oslo_vmware.api [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Task: {'id': task-1252920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249002} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.972709] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.974918] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 854.974918] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 854.974918] env[61936]: INFO nova.compute.manager [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Took 1.19 seconds to destroy the instance on the hypervisor. [ 854.974918] env[61936]: DEBUG oslo.service.loopingcall [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.974918] env[61936]: DEBUG nova.compute.manager [-] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 854.974918] env[61936]: DEBUG nova.network.neutron [-] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.998750] env[61936]: DEBUG nova.scheduler.client.report [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.122310] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252921, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061082} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.122597] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.123417] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15b31fb-d032-4ddb-b057-5b6a8ce82533 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.148074] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] f2844173-4114-4158-99ad-ac651efbb607/f2844173-4114-4158-99ad-ac651efbb607.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.150519] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64859d01-4b5d-4806-a946-5aef72d4aef3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.172907] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 855.172907] env[61936]: value = "task-1252926" [ 855.172907] env[61936]: _type = "Task" [ 855.172907] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.181451] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252926, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.240594] env[61936]: DEBUG nova.network.neutron [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.336825] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252922, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.341033] env[61936]: DEBUG oslo_concurrency.lockutils [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.392718] env[61936]: DEBUG nova.compute.manager [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received event network-vif-plugged-12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 855.392933] env[61936]: DEBUG oslo_concurrency.lockutils [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.393141] env[61936]: DEBUG oslo_concurrency.lockutils [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.393292] env[61936]: DEBUG oslo_concurrency.lockutils [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.393443] env[61936]: DEBUG nova.compute.manager [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] No waiting events found dispatching network-vif-plugged-12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 855.393596] env[61936]: WARNING nova.compute.manager [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received unexpected event network-vif-plugged-12397fd2-b04a-4019-b4fb-ca0f7b7fc152 for instance with vm_state building and task_state spawning. [ 855.393742] env[61936]: DEBUG nova.compute.manager [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received event network-changed-12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 855.393882] env[61936]: DEBUG nova.compute.manager [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Refreshing instance network info cache due to event network-changed-12397fd2-b04a-4019-b4fb-ca0f7b7fc152. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 855.394163] env[61936]: DEBUG oslo_concurrency.lockutils [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] Acquiring lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.426145] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b8092b-6105-c882-90af-696113520bd6, 'name': SearchDatastore_Task, 'duration_secs': 0.012278} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.426631] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.426922] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.427215] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e719316e-1635-4daf-8249-9b349aec32f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.434365] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 855.434365] env[61936]: value = "task-1252927" [ 855.434365] env[61936]: _type = "Task" [ 855.434365] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.444780] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252927, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.505628] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.506246] env[61936]: DEBUG nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 855.509441] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.046s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.509716] env[61936]: DEBUG nova.objects.instance [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lazy-loading 'resources' on Instance uuid d410e029-cb32-4cc6-b11d-9cba643e3b7b {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.570990] env[61936]: DEBUG nova.network.neutron [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.688594] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.721517] env[61936]: DEBUG nova.network.neutron [-] [instance: bdea3742-5789-4e97-b139-33cfee134843] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.837877] env[61936]: DEBUG oslo_vmware.api [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252922, 'name': PowerOnVM_Task, 'duration_secs': 0.948974} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.838305] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.838637] env[61936]: INFO nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Took 10.46 seconds to spawn the instance on the hypervisor. [ 855.838916] env[61936]: DEBUG nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 855.839961] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185a90ce-c8a7-4118-b036-01a400ec9946 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.948876] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252927, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.013955] env[61936]: DEBUG nova.compute.utils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.020086] env[61936]: DEBUG nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 856.020086] env[61936]: DEBUG nova.network.neutron [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 856.050969] env[61936]: DEBUG nova.compute.manager [req-e670c3ba-31d1-477f-b762-803572f1e342 req-9ba5e702-7605-426b-954b-1f893d367c53 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Received event network-vif-deleted-fc749785-35a5-468f-b3a3-b28dd3392c1c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 856.051230] env[61936]: INFO nova.compute.manager [req-e670c3ba-31d1-477f-b762-803572f1e342 req-9ba5e702-7605-426b-954b-1f893d367c53 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Neutron deleted interface fc749785-35a5-468f-b3a3-b28dd3392c1c; detaching it from the instance and deleting it from the info cache [ 856.051436] env[61936]: DEBUG nova.network.neutron [req-e670c3ba-31d1-477f-b762-803572f1e342 req-9ba5e702-7605-426b-954b-1f893d367c53 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.073908] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.074067] env[61936]: DEBUG nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Instance network_info: |[{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 856.077869] env[61936]: DEBUG oslo_concurrency.lockutils [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] Acquired lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.077869] env[61936]: DEBUG nova.network.neutron [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Refreshing network info cache for port 12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 856.077869] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:ca:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12397fd2-b04a-4019-b4fb-ca0f7b7fc152', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.083483] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Creating folder: Project (b9b0b8e03e6540a48947eebb51ad1135). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.088151] env[61936]: DEBUG nova.policy [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b98c1bbe1044663830b18c4ce9e23d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1b7d01ffac74c6fa2306b4839041fdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 856.091281] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0de0121e-45ab-4443-bb37-430536efd5ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.103047] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Created folder: Project (b9b0b8e03e6540a48947eebb51ad1135) in parent group-v269874. [ 856.103404] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Creating folder: Instances. Parent ref: group-v269939. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.103663] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82dd7346-4d0e-4140-9731-73215bd43644 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.116207] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Created folder: Instances in parent group-v269939. [ 856.116450] env[61936]: DEBUG oslo.service.loopingcall [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.119448] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.119661] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-064bbbe6-b6e0-4141-9892-bf39dc24e8db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.148577] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.148577] env[61936]: value = "task-1252930" [ 856.148577] env[61936]: _type = "Task" [ 856.148577] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.156125] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252930, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.183121] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252926, 'name': ReconfigVM_Task, 'duration_secs': 0.658748} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.185814] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Reconfigured VM instance instance-0000003d to attach disk [datastore1] f2844173-4114-4158-99ad-ac651efbb607/f2844173-4114-4158-99ad-ac651efbb607.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.186903] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bde67457-824b-4d81-b86e-cccc7756e50f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.193974] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 856.193974] env[61936]: value = "task-1252931" [ 856.193974] env[61936]: _type = "Task" [ 856.193974] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.203264] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252931, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.230503] env[61936]: INFO nova.compute.manager [-] [instance: bdea3742-5789-4e97-b139-33cfee134843] Took 1.29 seconds to deallocate network for instance. [ 856.269273] env[61936]: DEBUG nova.network.neutron [-] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.368207] env[61936]: INFO nova.compute.manager [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Took 41.05 seconds to build instance. [ 856.426966] env[61936]: DEBUG oslo_concurrency.lockutils [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.426966] env[61936]: DEBUG oslo_concurrency.lockutils [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.426966] env[61936]: INFO nova.compute.manager [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Attaching volume 56048ceb-814a-4561-9098-d35c809208f0 to /dev/sdb [ 856.448542] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fc4757-cac9-4c42-95a7-1a6ba82a3625 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.451227] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252927, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627541} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.451500] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.451709] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.452314] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28b6056b-fdd2-49c8-9e27-22cb2526b896 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.457417] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff67956-eaec-4a65-9128-0f115160e8a1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.462281] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 856.462281] env[61936]: value = "task-1252932" [ 856.462281] env[61936]: _type = "Task" [ 856.462281] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.493140] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c73ffc-c6bb-405a-82e6-340e60e56067 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.497466] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5626fb9d-0552-464e-afa7-3214a32689a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.504214] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252932, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.513016] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76cf17a-adb7-4fd6-bd9e-7d293c717a47 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.516724] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612024a0-625f-4f64-adf7-6148e4ee62c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.523759] env[61936]: DEBUG nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 856.540489] env[61936]: DEBUG nova.compute.provider_tree [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.545761] env[61936]: DEBUG nova.virt.block_device [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updating existing volume attachment record: af436c7c-586a-4da6-9f9e-5cc94ea7bd1a {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 856.554293] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-564771f2-d3b4-426c-a420-1559e5079e9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.564876] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc772a0-b911-474a-a0dc-30991ff5855b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.578624] env[61936]: DEBUG nova.network.neutron [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Successfully created port: b7963334-7c72-41c3-9e98-72a837dbb33f {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.594599] env[61936]: DEBUG nova.compute.manager [req-e670c3ba-31d1-477f-b762-803572f1e342 req-9ba5e702-7605-426b-954b-1f893d367c53 service nova] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Detach interface failed, port_id=fc749785-35a5-468f-b3a3-b28dd3392c1c, reason: Instance 301b0504-5cf9-44e0-bd3e-342f8a89278b could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 856.658576] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252930, 'name': CreateVM_Task, 'duration_secs': 0.378572} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.658755] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.659468] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.659644] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.659990] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.660262] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f14bd18-27ec-426c-9dc9-87f28e364adf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.664534] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 856.664534] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e26958-3ea7-88e2-778a-1655b70b53d5" [ 856.664534] env[61936]: _type = "Task" [ 856.664534] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.672180] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e26958-3ea7-88e2-778a-1655b70b53d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.703907] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252931, 'name': Rename_Task, 'duration_secs': 0.325388} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.704193] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.704437] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b96a91a-19ce-4a2a-addc-c7c8584604fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.711983] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 856.711983] env[61936]: value = "task-1252934" [ 856.711983] env[61936]: _type = "Task" [ 856.711983] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.720489] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252934, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.739607] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.772363] env[61936]: INFO nova.compute.manager [-] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Took 1.80 seconds to deallocate network for instance. [ 856.883277] env[61936]: DEBUG oslo_concurrency.lockutils [None req-71edb1cf-2566-4d0a-b0e8-27f89d1d1105 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.551s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.973790] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252932, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077117} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.976841] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.976841] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f612eb7f-9f29-484f-b1a0-765d9f5674ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.015193] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.017921] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58bfd957-72eb-452b-b3d5-60797f6ca7ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.042300] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 857.042300] env[61936]: value = "task-1252937" [ 857.042300] env[61936]: _type = "Task" [ 857.042300] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.051725] env[61936]: DEBUG nova.scheduler.client.report [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.054967] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252937, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.178555] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e26958-3ea7-88e2-778a-1655b70b53d5, 'name': SearchDatastore_Task, 'duration_secs': 0.047006} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.178555] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.178555] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.178555] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.178756] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.178756] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.178756] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05a63ae9-dac7-47cf-81d7-3ed611e2dd2c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.190021] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.192470] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.193035] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2782093b-be1f-4916-bad4-0cb129a43bd4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.198442] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 857.198442] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52898600-4373-9065-8304-84c7e2a793cb" [ 857.198442] env[61936]: _type = "Task" [ 857.198442] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.210451] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52898600-4373-9065-8304-84c7e2a793cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.223516] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252934, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.283676] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.295890] env[61936]: DEBUG nova.network.neutron [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updated VIF entry in instance network info cache for port 12397fd2-b04a-4019-b4fb-ca0f7b7fc152. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.296427] env[61936]: DEBUG nova.network.neutron [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.385296] env[61936]: DEBUG nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 857.540957] env[61936]: DEBUG nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 857.553388] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252937, 'name': ReconfigVM_Task, 'duration_secs': 0.44426} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.553707] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Reconfigured VM instance instance-0000003e to attach disk [datastore1] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.554395] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e44f23d-4a2c-42a1-b48c-7c427910aa54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.559805] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.050s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.563239] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.162s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.565658] env[61936]: INFO nova.compute.claims [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.571946] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 857.571946] env[61936]: value = "task-1252938" [ 857.571946] env[61936]: _type = "Task" [ 857.571946] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.576570] env[61936]: DEBUG nova.compute.manager [req-5aaebd97-d6a4-4898-b2a1-cfdc3bfe70f0 req-568f4b03-7260-405e-bc94-ac319dcdf1ed service nova] [instance: bdea3742-5789-4e97-b139-33cfee134843] Received event network-vif-deleted-74d05192-a7d3-48d2-b0e1-62141112c67e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 857.580748] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252938, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.583251] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 857.583409] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.583456] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 857.583631] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.583837] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 857.585159] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 857.585159] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 857.585159] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 857.585159] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 857.585159] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 857.585368] env[61936]: DEBUG nova.virt.hardware [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 857.587144] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9e510d-eb61-4b78-afb1-b842f7c5b8fa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.598143] env[61936]: INFO nova.scheduler.client.report [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted allocations for instance d410e029-cb32-4cc6-b11d-9cba643e3b7b [ 857.600702] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233b41bc-0e69-4a44-bdf3-67f8f84bb7bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.712271] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52898600-4373-9065-8304-84c7e2a793cb, 'name': SearchDatastore_Task, 'duration_secs': 0.022328} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.713304] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bcb8bc6-a9ac-40e5-966c-167cf4571828 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.726524] env[61936]: DEBUG oslo_vmware.api [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1252934, 'name': PowerOnVM_Task, 'duration_secs': 0.763234} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.728143] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.728866] env[61936]: INFO nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Took 9.88 seconds to spawn the instance on the hypervisor. [ 857.728866] env[61936]: DEBUG nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 857.729094] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 857.729094] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c3f86d-7604-67b8-9492-3b069462e0c9" [ 857.729094] env[61936]: _type = "Task" [ 857.729094] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.729865] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa67aec2-7100-4d0e-ad5f-162c171f0816 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.753461] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c3f86d-7604-67b8-9492-3b069462e0c9, 'name': SearchDatastore_Task, 'duration_secs': 0.012929} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.754373] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.754979] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] e15be950-1c32-4801-b4e4-a4d31ff0cfec/e15be950-1c32-4801-b4e4-a4d31ff0cfec.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.755505] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e31da82-0dd3-41c1-8627-d9a2ed10754b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.768644] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 857.768644] env[61936]: value = "task-1252939" [ 857.768644] env[61936]: _type = "Task" [ 857.768644] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.783574] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.801615] env[61936]: DEBUG oslo_concurrency.lockutils [req-aaf783e2-3b85-4856-b02e-890d5a0121fc req-035d6e82-025b-4133-9ef9-b6ab3fbeb30f service nova] Releasing lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.906796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.099221] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252938, 'name': Rename_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.116070] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7887f96b-7ae3-4555-9e5d-8abd32c01d7e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "d410e029-cb32-4cc6-b11d-9cba643e3b7b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.999s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.265277] env[61936]: INFO nova.compute.manager [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Took 42.38 seconds to build instance. [ 858.279143] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.430702] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86ac483-8c19-4417-bacc-69e3fb15fd9c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.443639] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e98cd98-d360-4120-bd00-a3716e5f8e83 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.487325] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5daa0ab1-68f5-43af-9d64-efcc2f07409b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.495642] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9c9b77-bfbf-47c9-b545-3ffa0e05f698 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.509837] env[61936]: DEBUG nova.compute.provider_tree [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.586842] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252938, 'name': Rename_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.660884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "e5e3e7af-8104-4f7a-ae18-81e03a932608" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.660884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.660884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "e5e3e7af-8104-4f7a-ae18-81e03a932608-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.660884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.661159] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.663493] env[61936]: INFO nova.compute.manager [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Terminating instance [ 858.769021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-07d374d8-8ebc-4d67-b6a9-680b98c780d1 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "f2844173-4114-4158-99ad-ac651efbb607" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.530s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.783025] env[61936]: DEBUG nova.compute.manager [req-79ef183d-266a-4b70-b78d-3e93da5f2985 req-acde5d64-4be8-466b-b144-c90726e53f13 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Received event network-vif-plugged-b7963334-7c72-41c3-9e98-72a837dbb33f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 858.784064] env[61936]: DEBUG oslo_concurrency.lockutils [req-79ef183d-266a-4b70-b78d-3e93da5f2985 req-acde5d64-4be8-466b-b144-c90726e53f13 service nova] Acquiring lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.784064] env[61936]: DEBUG oslo_concurrency.lockutils [req-79ef183d-266a-4b70-b78d-3e93da5f2985 req-acde5d64-4be8-466b-b144-c90726e53f13 service nova] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.784064] env[61936]: DEBUG oslo_concurrency.lockutils [req-79ef183d-266a-4b70-b78d-3e93da5f2985 req-acde5d64-4be8-466b-b144-c90726e53f13 service nova] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.784064] env[61936]: DEBUG nova.compute.manager [req-79ef183d-266a-4b70-b78d-3e93da5f2985 req-acde5d64-4be8-466b-b144-c90726e53f13 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] No waiting events found dispatching network-vif-plugged-b7963334-7c72-41c3-9e98-72a837dbb33f {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 858.784064] env[61936]: WARNING nova.compute.manager [req-79ef183d-266a-4b70-b78d-3e93da5f2985 req-acde5d64-4be8-466b-b144-c90726e53f13 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Received unexpected event network-vif-plugged-b7963334-7c72-41c3-9e98-72a837dbb33f for instance with vm_state building and task_state spawning. [ 858.792143] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.002683] env[61936]: DEBUG nova.network.neutron [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Successfully updated port: b7963334-7c72-41c3-9e98-72a837dbb33f {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.016598] env[61936]: DEBUG nova.scheduler.client.report [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 859.087248] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252938, 'name': Rename_Task, 'duration_secs': 1.160616} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.087602] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 859.087894] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c148964-32b7-4b9b-a06e-207f2662d2f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.096939] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 859.096939] env[61936]: value = "task-1252942" [ 859.096939] env[61936]: _type = "Task" [ 859.096939] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.110354] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252942, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.167280] env[61936]: DEBUG nova.compute.manager [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 859.168265] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.169759] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8863d57-dcad-4598-b3fe-b342c12ed40c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.179999] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.180498] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe9b5197-25e8-4cb9-98a6-b2179c7dd618 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.188723] env[61936]: DEBUG oslo_vmware.api [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 859.188723] env[61936]: value = "task-1252943" [ 859.188723] env[61936]: _type = "Task" [ 859.188723] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.200318] env[61936]: DEBUG oslo_vmware.api [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252943, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.270720] env[61936]: DEBUG nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 859.290231] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252939, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.510723] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "refresh_cache-18175943-35d2-4792-a8ea-e54b4ffbaf2f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.510887] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "refresh_cache-18175943-35d2-4792-a8ea-e54b4ffbaf2f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.511074] env[61936]: DEBUG nova.network.neutron [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.524117] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.961s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.524882] env[61936]: DEBUG nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 859.532029] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.493s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.532029] env[61936]: INFO nova.compute.claims [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.608873] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252942, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.700519] env[61936]: DEBUG oslo_vmware.api [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252943, 'name': PowerOffVM_Task, 'duration_secs': 0.479534} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.700815] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.703739] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.703739] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-894d911a-03c2-474d-8b56-23a42ae9531b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.786336] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.786515] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.786729] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Deleting the datastore file [datastore2] e5e3e7af-8104-4f7a-ae18-81e03a932608 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.789077] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df4d66e1-a814-4992-9f04-5beaa9b7d135 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.800629] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252939, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.563645} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.801857] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] e15be950-1c32-4801-b4e4-a4d31ff0cfec/e15be950-1c32-4801-b4e4-a4d31ff0cfec.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.801857] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.801978] env[61936]: DEBUG oslo_vmware.api [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for the task: (returnval){ [ 859.801978] env[61936]: value = "task-1252945" [ 859.801978] env[61936]: _type = "Task" [ 859.801978] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.802957] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.803222] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb321533-0d54-4666-b975-023eb281e9db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.815126] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 859.815126] env[61936]: value = "task-1252946" [ 859.815126] env[61936]: _type = "Task" [ 859.815126] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.823962] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.002520] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.002773] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.036488] env[61936]: DEBUG nova.compute.utils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.040108] env[61936]: DEBUG nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 860.040332] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 860.083510] env[61936]: DEBUG nova.network.neutron [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.114796] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252942, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.169780] env[61936]: DEBUG nova.policy [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3382c0703a24837a7a62574bfc6ed37', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f03f2cf1d4084fdca2664ccaac0a7671', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.317836] env[61936]: DEBUG oslo_vmware.api [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Task: {'id': task-1252945, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.456727} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.321480] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.321632] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.321851] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.322040] env[61936]: INFO nova.compute.manager [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Took 1.15 seconds to destroy the instance on the hypervisor. [ 860.322285] env[61936]: DEBUG oslo.service.loopingcall [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.322486] env[61936]: DEBUG nova.compute.manager [-] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 860.322581] env[61936]: DEBUG nova.network.neutron [-] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.331732] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068631} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.331844] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.332685] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45cc449-6613-4f2b-a2cd-a05b882212ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.357354] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] e15be950-1c32-4801-b4e4-a4d31ff0cfec/e15be950-1c32-4801-b4e4-a4d31ff0cfec.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.358026] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c2a243c-0b4f-4e7f-b1f7-1c8ecb504e85 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.383568] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 860.383568] env[61936]: value = "task-1252948" [ 860.383568] env[61936]: _type = "Task" [ 860.383568] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.391834] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252948, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.491590] env[61936]: DEBUG nova.network.neutron [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Updating instance_info_cache with network_info: [{"id": "b7963334-7c72-41c3-9e98-72a837dbb33f", "address": "fa:16:3e:45:60:cf", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7963334-7c", "ovs_interfaceid": "b7963334-7c72-41c3-9e98-72a837dbb33f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.544584] env[61936]: DEBUG nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 860.616123] env[61936]: DEBUG oslo_vmware.api [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252942, 'name': PowerOnVM_Task, 'duration_secs': 1.189405} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.616442] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.618066] env[61936]: INFO nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Took 10.22 seconds to spawn the instance on the hypervisor. [ 860.618066] env[61936]: DEBUG nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 860.618066] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2b547c-71d1-4a72-8bdb-c321a601b30f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.802984] env[61936]: DEBUG nova.compute.manager [req-f05cea56-f318-4f4c-80d6-662d813009b6 req-d6a8a2f8-5918-4fa8-80cb-433d01d65399 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Received event network-vif-deleted-ca6d9c8e-18c6-4a80-a792-baa44a0acc54 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 860.803181] env[61936]: INFO nova.compute.manager [req-f05cea56-f318-4f4c-80d6-662d813009b6 req-d6a8a2f8-5918-4fa8-80cb-433d01d65399 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Neutron deleted interface ca6d9c8e-18c6-4a80-a792-baa44a0acc54; detaching it from the instance and deleting it from the info cache [ 860.803352] env[61936]: DEBUG nova.network.neutron [req-f05cea56-f318-4f4c-80d6-662d813009b6 req-d6a8a2f8-5918-4fa8-80cb-433d01d65399 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.814016] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Successfully created port: 180a6391-4494-456c-b805-ec2ad1617c73 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.890952] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b598e96f-7569-44e2-8eb9-bdaa9c8e008a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.896545] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252948, 'name': ReconfigVM_Task, 'duration_secs': 0.286226} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.898139] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Reconfigured VM instance instance-0000003f to attach disk [datastore1] e15be950-1c32-4801-b4e4-a4d31ff0cfec/e15be950-1c32-4801-b4e4-a4d31ff0cfec.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.898818] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61a1d007-6643-490d-8743-18bbe55340aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.903128] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe47e253-b3dc-4fa5-bca4-db6ff7a94e92 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.907939] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 860.907939] env[61936]: value = "task-1252949" [ 860.907939] env[61936]: _type = "Task" [ 860.907939] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.939414] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac569bcc-4cca-4273-85a0-52851f2ad3b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.945499] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252949, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.951882] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a5b792-acbd-42c0-8b48-737437dc039b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.966638] env[61936]: DEBUG nova.compute.provider_tree [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.994516] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "refresh_cache-18175943-35d2-4792-a8ea-e54b4ffbaf2f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.994877] env[61936]: DEBUG nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Instance network_info: |[{"id": "b7963334-7c72-41c3-9e98-72a837dbb33f", "address": "fa:16:3e:45:60:cf", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7963334-7c", "ovs_interfaceid": "b7963334-7c72-41c3-9e98-72a837dbb33f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 860.995555] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:60:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7963334-7c72-41c3-9e98-72a837dbb33f', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.004059] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating folder: Project (e1b7d01ffac74c6fa2306b4839041fdd). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.004672] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-13024e64-2f44-4276-b526-3e2227d63153 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.017581] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Created folder: Project (e1b7d01ffac74c6fa2306b4839041fdd) in parent group-v269874. [ 861.017807] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating folder: Instances. Parent ref: group-v269944. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.018131] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bfb8c58-f452-4d69-9eb9-fb1f5ac86e9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.029179] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Created folder: Instances in parent group-v269944. [ 861.029559] env[61936]: DEBUG oslo.service.loopingcall [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.029842] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.030099] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0a71f74-15e0-4419-b5c3-326bcfa12540 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.054175] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.054175] env[61936]: value = "task-1252952" [ 861.054175] env[61936]: _type = "Task" [ 861.054175] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.069956] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252952, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.071691] env[61936]: DEBUG nova.compute.manager [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Received event network-changed-b7963334-7c72-41c3-9e98-72a837dbb33f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 861.071873] env[61936]: DEBUG nova.compute.manager [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Refreshing instance network info cache due to event network-changed-b7963334-7c72-41c3-9e98-72a837dbb33f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 861.072092] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Acquiring lock "refresh_cache-18175943-35d2-4792-a8ea-e54b4ffbaf2f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.072243] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Acquired lock "refresh_cache-18175943-35d2-4792-a8ea-e54b4ffbaf2f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.072417] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Refreshing network info cache for port b7963334-7c72-41c3-9e98-72a837dbb33f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.100716] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 861.100716] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269943', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'name': 'volume-56048ceb-814a-4561-9098-d35c809208f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377', 'attached_at': '', 'detached_at': '', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'serial': '56048ceb-814a-4561-9098-d35c809208f0'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 861.102023] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68718c70-5080-4856-865b-7906bbde0ef0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.121917] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b32b1c4-27a3-44e1-9350-c62b31b4f6d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.142436] env[61936]: DEBUG nova.network.neutron [-] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.151877] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] volume-56048ceb-814a-4561-9098-d35c809208f0/volume-56048ceb-814a-4561-9098-d35c809208f0.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.156271] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a056e00e-9108-430f-8a8a-8da590f9af65 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.178021] env[61936]: INFO nova.compute.manager [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Took 42.36 seconds to build instance. [ 861.183910] env[61936]: DEBUG oslo_vmware.api [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 861.183910] env[61936]: value = "task-1252953" [ 861.183910] env[61936]: _type = "Task" [ 861.183910] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.194396] env[61936]: DEBUG oslo_vmware.api [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252953, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.307357] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3628a829-740d-40c7-95bc-747abfd3931a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.318040] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed761be-bf26-4ea1-a5c4-5d4b8360fe43 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.349750] env[61936]: DEBUG nova.compute.manager [req-f05cea56-f318-4f4c-80d6-662d813009b6 req-d6a8a2f8-5918-4fa8-80cb-433d01d65399 service nova] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Detach interface failed, port_id=ca6d9c8e-18c6-4a80-a792-baa44a0acc54, reason: Instance e5e3e7af-8104-4f7a-ae18-81e03a932608 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 861.383312] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Successfully created port: cf56feac-af85-4393-84a6-d364545347cc {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.418289] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252949, 'name': Rename_Task, 'duration_secs': 0.179763} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.418569] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.418816] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af006cf3-76c3-4a94-add7-2e58390c4007 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.425579] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 861.425579] env[61936]: value = "task-1252954" [ 861.425579] env[61936]: _type = "Task" [ 861.425579] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.433761] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.470992] env[61936]: DEBUG nova.scheduler.client.report [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 861.568859] env[61936]: DEBUG nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 861.570980] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252952, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.597652] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 861.597897] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.598061] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 861.598311] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.598497] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 861.598657] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 861.598874] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 861.599032] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 861.599246] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 861.599353] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 861.599519] env[61936]: DEBUG nova.virt.hardware [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 861.600548] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7da2c8-08e8-4797-8ed6-7b5c33fd5d0f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.610103] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ac0f5a-d33a-48e2-807f-5f040dcdbba7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.652975] env[61936]: INFO nova.compute.manager [-] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Took 1.33 seconds to deallocate network for instance. [ 861.680359] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6600eca8-a995-4c91-8c6a-c715d6993eee tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "e800f5e9-deb0-4592-aded-3768d238e311" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.080s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.694870] env[61936]: DEBUG oslo_vmware.api [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.829136] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Updated VIF entry in instance network info cache for port b7963334-7c72-41c3-9e98-72a837dbb33f. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.829555] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Updating instance_info_cache with network_info: [{"id": "b7963334-7c72-41c3-9e98-72a837dbb33f", "address": "fa:16:3e:45:60:cf", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7963334-7c", "ovs_interfaceid": "b7963334-7c72-41c3-9e98-72a837dbb33f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.936655] env[61936]: DEBUG oslo_vmware.api [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1252954, 'name': PowerOnVM_Task, 'duration_secs': 0.482201} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.938151] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.938151] env[61936]: INFO nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Took 9.01 seconds to spawn the instance on the hypervisor. [ 861.938151] env[61936]: DEBUG nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 861.938151] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429e419e-6169-4b43-b230-36ce7affacfb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.977675] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.978246] env[61936]: DEBUG nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 861.981584] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.866s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.984912] env[61936]: INFO nova.compute.claims [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.065830] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252952, 'name': CreateVM_Task, 'duration_secs': 0.55118} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.066115] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.066735] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.066914] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.067246] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.067508] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9283947b-6419-4aa7-a295-c9bd1dd71926 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.074050] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 862.074050] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a48f64-58b5-6a04-6100-4173dbb162ea" [ 862.074050] env[61936]: _type = "Task" [ 862.074050] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.082125] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a48f64-58b5-6a04-6100-4173dbb162ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.160172] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.184583] env[61936]: DEBUG nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 862.198372] env[61936]: DEBUG oslo_vmware.api [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252953, 'name': ReconfigVM_Task, 'duration_secs': 0.627944} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.198709] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Reconfigured VM instance instance-00000035 to attach disk [datastore2] volume-56048ceb-814a-4561-9098-d35c809208f0/volume-56048ceb-814a-4561-9098-d35c809208f0.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.204356] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-662a0911-63de-41d2-b0d4-8e5547941926 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.221106] env[61936]: DEBUG oslo_vmware.api [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 862.221106] env[61936]: value = "task-1252956" [ 862.221106] env[61936]: _type = "Task" [ 862.221106] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.230216] env[61936]: DEBUG oslo_vmware.api [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252956, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.332821] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Releasing lock "refresh_cache-18175943-35d2-4792-a8ea-e54b4ffbaf2f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.333193] env[61936]: DEBUG nova.compute.manager [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Received event network-changed-dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 862.333319] env[61936]: DEBUG nova.compute.manager [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Refreshing instance network info cache due to event network-changed-dd14ed80-5888-44af-897b-0ed4dd854e0d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 862.333546] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Acquiring lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.334094] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Acquired lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.334094] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Refreshing network info cache for port dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.461035] env[61936]: INFO nova.compute.manager [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Took 40.09 seconds to build instance. [ 862.487241] env[61936]: DEBUG nova.compute.utils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.490660] env[61936]: DEBUG nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Not allocating networking since 'none' was specified. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 862.586360] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a48f64-58b5-6a04-6100-4173dbb162ea, 'name': SearchDatastore_Task, 'duration_secs': 0.011675} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.586803] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.587145] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.587366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.587570] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.587853] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.588250] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e92c07e-577a-48c2-a639-7b53070fadde {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.599156] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.599373] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 862.600135] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc012b3c-fa44-42cc-af72-76af520c5472 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.606696] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 862.606696] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cc4b99-2f18-4e4f-4c0a-413fd786e912" [ 862.606696] env[61936]: _type = "Task" [ 862.606696] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.614900] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cc4b99-2f18-4e4f-4c0a-413fd786e912, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.714451] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.731696] env[61936]: DEBUG oslo_vmware.api [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1252956, 'name': ReconfigVM_Task, 'duration_secs': 0.147577} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.732015] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269943', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'name': 'volume-56048ceb-814a-4561-9098-d35c809208f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377', 'attached_at': '', 'detached_at': '', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'serial': '56048ceb-814a-4561-9098-d35c809208f0'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 862.963546] env[61936]: DEBUG oslo_concurrency.lockutils [None req-91e5caf4-fee6-4492-b131-0b063e0b498a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.045s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.995309] env[61936]: DEBUG nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 863.057024] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Updated VIF entry in instance network info cache for port dd14ed80-5888-44af-897b-0ed4dd854e0d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.057024] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Updating instance_info_cache with network_info: [{"id": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "address": "fa:16:3e:87:2f:dd", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd14ed80-58", "ovs_interfaceid": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.118447] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cc4b99-2f18-4e4f-4c0a-413fd786e912, 'name': SearchDatastore_Task, 'duration_secs': 0.025904} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.119273] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1adec48-6910-4223-9f8a-b6ffe4e6cc57 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.124886] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 863.124886] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520e09dc-b7de-9594-8173-e4ac2d58134a" [ 863.124886] env[61936]: _type = "Task" [ 863.124886] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.136724] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520e09dc-b7de-9594-8173-e4ac2d58134a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.311838] env[61936]: INFO nova.compute.manager [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Rebuilding instance [ 863.353276] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a689a5-ee77-4bc6-95d3-d870baabb086 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.362271] env[61936]: DEBUG nova.compute.manager [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 863.363082] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551c8a46-4cb4-473b-b6b4-95f8f89a6afb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.368844] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961f48af-c62d-430d-8319-f310e6ad351d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.404554] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2de877d-aef0-4f01-809f-8baf70006882 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.416077] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e381573d-9bc5-4c3f-b1a3-82107c975e08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.429437] env[61936]: DEBUG nova.compute.provider_tree [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.466404] env[61936]: DEBUG nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 863.559502] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Releasing lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.559773] env[61936]: DEBUG nova.compute.manager [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Received event network-changed-dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 863.559938] env[61936]: DEBUG nova.compute.manager [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Refreshing instance network info cache due to event network-changed-dd14ed80-5888-44af-897b-0ed4dd854e0d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 863.560218] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Acquiring lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.560394] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Acquired lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.560632] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Refreshing network info cache for port dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.636074] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520e09dc-b7de-9594-8173-e4ac2d58134a, 'name': SearchDatastore_Task, 'duration_secs': 0.0382} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.636347] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.636595] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 18175943-35d2-4792-a8ea-e54b4ffbaf2f/18175943-35d2-4792-a8ea-e54b4ffbaf2f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.636845] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-98cf15d2-35d1-4728-a8fe-095d39cf905c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.644817] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 863.644817] env[61936]: value = "task-1252957" [ 863.644817] env[61936]: _type = "Task" [ 863.644817] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.653339] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.781298] env[61936]: DEBUG nova.objects.instance [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'flavor' on Instance uuid 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.845526] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Successfully updated port: 180a6391-4494-456c-b805-ec2ad1617c73 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.932189] env[61936]: DEBUG nova.scheduler.client.report [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 863.975613] env[61936]: DEBUG nova.compute.manager [req-05d1aa60-d2a2-4c70-bef4-4b77470c7854 req-a6d9c1e6-ee86-4ffd-ae34-0b7f5404c1df service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received event network-vif-plugged-180a6391-4494-456c-b805-ec2ad1617c73 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 863.975951] env[61936]: DEBUG oslo_concurrency.lockutils [req-05d1aa60-d2a2-4c70-bef4-4b77470c7854 req-a6d9c1e6-ee86-4ffd-ae34-0b7f5404c1df service nova] Acquiring lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.976137] env[61936]: DEBUG oslo_concurrency.lockutils [req-05d1aa60-d2a2-4c70-bef4-4b77470c7854 req-a6d9c1e6-ee86-4ffd-ae34-0b7f5404c1df service nova] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.976397] env[61936]: DEBUG oslo_concurrency.lockutils [req-05d1aa60-d2a2-4c70-bef4-4b77470c7854 req-a6d9c1e6-ee86-4ffd-ae34-0b7f5404c1df service nova] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.976555] env[61936]: DEBUG nova.compute.manager [req-05d1aa60-d2a2-4c70-bef4-4b77470c7854 req-a6d9c1e6-ee86-4ffd-ae34-0b7f5404c1df service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] No waiting events found dispatching network-vif-plugged-180a6391-4494-456c-b805-ec2ad1617c73 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 863.976758] env[61936]: WARNING nova.compute.manager [req-05d1aa60-d2a2-4c70-bef4-4b77470c7854 req-a6d9c1e6-ee86-4ffd-ae34-0b7f5404c1df service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received unexpected event network-vif-plugged-180a6391-4494-456c-b805-ec2ad1617c73 for instance with vm_state building and task_state spawning. [ 863.993978] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.011021] env[61936]: DEBUG nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 864.045866] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 864.046156] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.046312] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 864.046489] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.046631] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 864.046772] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 864.046969] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 864.047195] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 864.047387] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 864.047533] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 864.047698] env[61936]: DEBUG nova.virt.hardware [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 864.049085] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9988eacc-779b-4aee-89c6-ea082c976332 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.061723] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214a0058-4d51-40e8-83dd-eda33692db99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.086918] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.092772] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Creating folder: Project (9173e087a76a4284915d63798f5345de). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.093832] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d6de757-89af-4d14-baee-8f665e7d47e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.107454] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Created folder: Project (9173e087a76a4284915d63798f5345de) in parent group-v269874. [ 864.107454] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Creating folder: Instances. Parent ref: group-v269947. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.107454] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f0db811b-31cf-4c28-8b5f-c28be14849b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.118646] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Created folder: Instances in parent group-v269947. [ 864.118646] env[61936]: DEBUG oslo.service.loopingcall [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.120063] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.120063] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-24daeca1-d563-43ba-9752-d9abc8ca268c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.145614] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.145614] env[61936]: value = "task-1252960" [ 864.145614] env[61936]: _type = "Task" [ 864.145614] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.158678] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252957, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.162423] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252960, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.288239] env[61936]: DEBUG oslo_concurrency.lockutils [None req-787ce43e-8846-4be6-9db8-6810412adfa8 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.863s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.382909] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.384185] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17ff8f22-6a48-46d0-9800-f9cc03d1762e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.394745] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 864.394745] env[61936]: value = "task-1252962" [ 864.394745] env[61936]: _type = "Task" [ 864.394745] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.405935] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252962, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.437124] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.437855] env[61936]: DEBUG nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 864.441611] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.466s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.442098] env[61936]: INFO nova.compute.claims [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.532836] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Updated VIF entry in instance network info cache for port dd14ed80-5888-44af-897b-0ed4dd854e0d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 864.532836] env[61936]: DEBUG nova.network.neutron [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Updating instance_info_cache with network_info: [{"id": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "address": "fa:16:3e:87:2f:dd", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd14ed80-58", "ovs_interfaceid": "dd14ed80-5888-44af-897b-0ed4dd854e0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.670067] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.872932} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.677450] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 18175943-35d2-4792-a8ea-e54b4ffbaf2f/18175943-35d2-4792-a8ea-e54b4ffbaf2f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.677889] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.678285] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252960, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.678640] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19717b4f-8cb1-4644-b56c-e1c27674a8b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.692821] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 864.692821] env[61936]: value = "task-1252963" [ 864.692821] env[61936]: _type = "Task" [ 864.692821] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.710240] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252963, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.911088] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252962, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.946840] env[61936]: DEBUG nova.compute.utils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.954625] env[61936]: DEBUG nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Not allocating networking since 'none' was specified. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 865.035569] env[61936]: DEBUG oslo_concurrency.lockutils [req-8fa152bb-6104-4928-ac74-3f28389beb53 req-23d209ee-ea9e-40c3-a822-532954c432f4 service nova] Releasing lock "refresh_cache-f2844173-4114-4158-99ad-ac651efbb607" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.167259] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252960, 'name': CreateVM_Task, 'duration_secs': 0.663918} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.167495] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.167973] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.168248] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.168654] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.168976] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad5c37cd-d88a-4a63-8f60-d83b6b7db426 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.174859] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 865.174859] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ad8605-e8cb-3060-2682-78cfdcf0bf72" [ 865.174859] env[61936]: _type = "Task" [ 865.174859] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.184827] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ad8605-e8cb-3060-2682-78cfdcf0bf72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.202817] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252963, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092758} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.203167] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.204153] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e68f26d-6457-4a78-bf30-b88c1c00a53c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.236150] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 18175943-35d2-4792-a8ea-e54b4ffbaf2f/18175943-35d2-4792-a8ea-e54b4ffbaf2f.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.236150] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-347b1a3d-6fc7-4f71-b7f1-2b930c4f286b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.262032] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 865.262032] env[61936]: value = "task-1252964" [ 865.262032] env[61936]: _type = "Task" [ 865.262032] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.268384] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.407109] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252962, 'name': PowerOffVM_Task, 'duration_secs': 0.546691} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.407654] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.408043] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.409027] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8264b67-5320-402d-958f-9debffa58f33 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.419017] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.419017] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71e2b49d-d22b-45f8-8dc4-f3472d404224 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.456082] env[61936]: DEBUG nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 865.503629] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.503629] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.503862] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleting the datastore file [datastore1] e800f5e9-deb0-4592-aded-3768d238e311 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.504059] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90e42ab2-4766-48af-8677-28fc9e810c53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.515153] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 865.515153] env[61936]: value = "task-1252966" [ 865.515153] env[61936]: _type = "Task" [ 865.515153] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.527487] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252966, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.686054] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ad8605-e8cb-3060-2682-78cfdcf0bf72, 'name': SearchDatastore_Task, 'duration_secs': 0.010107} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.689139] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.689469] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.689690] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.689849] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.690050] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.690537] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee5d5c3b-9ea2-4fc5-a2ba-083612691075 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.700384] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.700384] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.703434] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c792f43d-d638-4001-a289-e3df98d93393 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.711229] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 865.711229] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52480d1f-e277-a119-a646-bd9a542fa642" [ 865.711229] env[61936]: _type = "Task" [ 865.711229] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.718811] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52480d1f-e277-a119-a646-bd9a542fa642, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.779587] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252964, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.825640] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954340ae-085f-4a7c-8b95-cf3c26bae6e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.833806] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce2530e-9964-45e0-b52d-0862d7bd59b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.870067] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497af843-2f9f-4479-a8f0-c23ba3788f66 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.878656] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d63339-2870-4e3b-95cc-6ad2d5e2af6e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.892549] env[61936]: DEBUG nova.compute.provider_tree [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.971749] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Successfully updated port: cf56feac-af85-4393-84a6-d364545347cc {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.025451] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252966, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180077} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.025693] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.025907] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.026057] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.222593] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52480d1f-e277-a119-a646-bd9a542fa642, 'name': SearchDatastore_Task, 'duration_secs': 0.010156} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.223487] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d1d1249-4833-4de9-ac56-e7e8ea4981bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.229698] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 866.229698] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b4fbad-ba55-7745-5ffc-346cdde07003" [ 866.229698] env[61936]: _type = "Task" [ 866.229698] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.240407] env[61936]: DEBUG nova.compute.manager [req-b204be7f-3773-4504-967b-9c33730f360a req-53b6a4cc-e4b6-4edd-b5f0-9c7b13abb144 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received event network-vif-plugged-cf56feac-af85-4393-84a6-d364545347cc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 866.240640] env[61936]: DEBUG oslo_concurrency.lockutils [req-b204be7f-3773-4504-967b-9c33730f360a req-53b6a4cc-e4b6-4edd-b5f0-9c7b13abb144 service nova] Acquiring lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.240844] env[61936]: DEBUG oslo_concurrency.lockutils [req-b204be7f-3773-4504-967b-9c33730f360a req-53b6a4cc-e4b6-4edd-b5f0-9c7b13abb144 service nova] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.241013] env[61936]: DEBUG oslo_concurrency.lockutils [req-b204be7f-3773-4504-967b-9c33730f360a req-53b6a4cc-e4b6-4edd-b5f0-9c7b13abb144 service nova] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.241187] env[61936]: DEBUG nova.compute.manager [req-b204be7f-3773-4504-967b-9c33730f360a req-53b6a4cc-e4b6-4edd-b5f0-9c7b13abb144 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] No waiting events found dispatching network-vif-plugged-cf56feac-af85-4393-84a6-d364545347cc {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 866.241398] env[61936]: WARNING nova.compute.manager [req-b204be7f-3773-4504-967b-9c33730f360a req-53b6a4cc-e4b6-4edd-b5f0-9c7b13abb144 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received unexpected event network-vif-plugged-cf56feac-af85-4393-84a6-d364545347cc for instance with vm_state building and task_state spawning. [ 866.248117] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b4fbad-ba55-7745-5ffc-346cdde07003, 'name': SearchDatastore_Task, 'duration_secs': 0.010024} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.248397] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.248647] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 258fed0b-ba19-4c90-981a-96bb36ba8bc8/258fed0b-ba19-4c90-981a-96bb36ba8bc8.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.248909] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-444b3094-911a-4cd1-b216-3627b4dcede2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.256533] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 866.256533] env[61936]: value = "task-1252968" [ 866.256533] env[61936]: _type = "Task" [ 866.256533] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.276808] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252964, 'name': ReconfigVM_Task, 'duration_secs': 0.601928} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.280335] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 18175943-35d2-4792-a8ea-e54b4ffbaf2f/18175943-35d2-4792-a8ea-e54b4ffbaf2f.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.281565] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252968, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.281819] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6c9fc68-80ab-459b-95c5-7b9122c1b71e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.290509] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 866.290509] env[61936]: value = "task-1252969" [ 866.290509] env[61936]: _type = "Task" [ 866.290509] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.300489] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252969, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.319682] env[61936]: DEBUG nova.compute.manager [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received event network-changed-180a6391-4494-456c-b805-ec2ad1617c73 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 866.320575] env[61936]: DEBUG nova.compute.manager [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Refreshing instance network info cache due to event network-changed-180a6391-4494-456c-b805-ec2ad1617c73. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 866.320575] env[61936]: DEBUG oslo_concurrency.lockutils [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] Acquiring lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.320575] env[61936]: DEBUG oslo_concurrency.lockutils [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] Acquired lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.320758] env[61936]: DEBUG nova.network.neutron [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Refreshing network info cache for port 180a6391-4494-456c-b805-ec2ad1617c73 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.396014] env[61936]: DEBUG nova.scheduler.client.report [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 866.468233] env[61936]: DEBUG nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 866.474345] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.496436] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 866.497082] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.497082] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 866.497082] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.497298] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 866.497476] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 866.497621] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 866.497775] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 866.497934] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 866.498118] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 866.498338] env[61936]: DEBUG nova.virt.hardware [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 866.499232] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdabab01-36ae-46b2-b671-bab2f566690f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.512422] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7544ab8f-03d1-4292-9bbd-f5a045325126 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.535010] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.541286] env[61936]: DEBUG oslo.service.loopingcall [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.541936] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.542189] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e43b6bc-265b-4fd6-b25d-884887bd033b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.565931] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.565931] env[61936]: value = "task-1252970" [ 866.565931] env[61936]: _type = "Task" [ 866.565931] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.582351] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252970, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.759676] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "10669072-a8e8-4262-b390-8bf31af6fafe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.759853] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "10669072-a8e8-4262-b390-8bf31af6fafe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.770195] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252968, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503226} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.770509] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 258fed0b-ba19-4c90-981a-96bb36ba8bc8/258fed0b-ba19-4c90-981a-96bb36ba8bc8.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.770933] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.771038] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85fb1d96-2614-4ab7-a769-b55d809bc786 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.777981] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 866.777981] env[61936]: value = "task-1252971" [ 866.777981] env[61936]: _type = "Task" [ 866.777981] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.788769] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252971, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.800888] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252969, 'name': Rename_Task, 'duration_secs': 0.162699} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.801219] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.801469] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-440e41c5-2245-4a2e-9253-7feb00fe508a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.808766] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 866.808766] env[61936]: value = "task-1252972" [ 866.808766] env[61936]: _type = "Task" [ 866.808766] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.816942] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.821084] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "2797bce9-f221-49b5-ab2b-42df3f347497" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.821305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.859692] env[61936]: DEBUG nova.network.neutron [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.901494] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.902038] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 866.904617] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.341s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.905982] env[61936]: INFO nova.compute.claims [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.948695] env[61936]: DEBUG nova.network.neutron [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.071554] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 867.071823] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.071977] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 867.072170] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.072317] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 867.072462] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 867.072663] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 867.072818] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 867.073027] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 867.073203] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 867.073374] env[61936]: DEBUG nova.virt.hardware [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 867.074130] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffc86e1-4ffd-4708-9a22-93b9e1d162a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.085165] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d8a86e-d6da-48e2-8094-fd42d8dc7253 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.089066] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252970, 'name': CreateVM_Task, 'duration_secs': 0.395688} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.089698] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.090430] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.090600] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.090905] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.091142] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c093ecb-86f9-463b-832a-b090ac9074ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.100463] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:2f:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b94712a6-b777-47dd-bc06-f9acfce2d936', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48edeab7-9688-4106-9b4a-9e53a69aa280', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.108103] env[61936]: DEBUG oslo.service.loopingcall [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.108679] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.108879] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10989af9-5006-4e57-b805-f795a1eceec1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.124202] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 867.124202] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c1f409-76ff-931c-0509-93d2852d918a" [ 867.124202] env[61936]: _type = "Task" [ 867.124202] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.129358] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.129358] env[61936]: value = "task-1252973" [ 867.129358] env[61936]: _type = "Task" [ 867.129358] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.132786] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c1f409-76ff-931c-0509-93d2852d918a, 'name': SearchDatastore_Task, 'duration_secs': 0.010626} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.135737] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.135995] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.136243] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.136387] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.136557] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.136789] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66b3d916-7c9a-4eaa-ac50-427853e88eb4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.143104] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252973, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.144618] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.144792] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.145467] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d33e641-67a7-463e-bb8f-baf33a46059d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.150497] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 867.150497] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e726ad-95a6-0a6d-4d0c-ac1304139312" [ 867.150497] env[61936]: _type = "Task" [ 867.150497] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.158048] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e726ad-95a6-0a6d-4d0c-ac1304139312, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.288817] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252971, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089795} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.289171] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.289935] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016ddaa3-15ac-4e98-b8f3-f77b275d6c08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.310289] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 258fed0b-ba19-4c90-981a-96bb36ba8bc8/258fed0b-ba19-4c90-981a-96bb36ba8bc8.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.310601] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99d1d0b5-2b81-456c-acd8-f227e8b39ec0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.333605] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252972, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.334970] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 867.334970] env[61936]: value = "task-1252974" [ 867.334970] env[61936]: _type = "Task" [ 867.334970] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.342954] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252974, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.414059] env[61936]: DEBUG nova.compute.utils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.419116] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 867.419116] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 867.453773] env[61936]: DEBUG oslo_concurrency.lockutils [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] Releasing lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.453773] env[61936]: DEBUG nova.compute.manager [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received event network-changed-12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 867.453773] env[61936]: DEBUG nova.compute.manager [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Refreshing instance network info cache due to event network-changed-12397fd2-b04a-4019-b4fb-ca0f7b7fc152. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 867.453773] env[61936]: DEBUG oslo_concurrency.lockutils [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] Acquiring lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.453773] env[61936]: DEBUG oslo_concurrency.lockutils [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] Acquired lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.454291] env[61936]: DEBUG nova.network.neutron [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Refreshing network info cache for port 12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.454291] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquired lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.454291] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 867.472562] env[61936]: DEBUG nova.policy [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6aec9537a6bc46b6b55b3163f1b59cc5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91a49a1ba2c34c8080cef82a7355cd95', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.647028] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252973, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.664012] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e726ad-95a6-0a6d-4d0c-ac1304139312, 'name': SearchDatastore_Task, 'duration_secs': 0.0093} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.665193] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e43de58e-073a-47e0-bfd4-bc863c56c619 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.672230] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 867.672230] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526115c2-7c4b-1a81-12c9-e6b4aafd6502" [ 867.672230] env[61936]: _type = "Task" [ 867.672230] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.682553] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526115c2-7c4b-1a81-12c9-e6b4aafd6502, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.763015] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Successfully created port: dc5d4628-264f-4f4e-9006-834caeb7b139 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.821015] env[61936]: DEBUG oslo_vmware.api [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252972, 'name': PowerOnVM_Task, 'duration_secs': 0.873277} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.821295] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.821505] env[61936]: INFO nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Took 10.28 seconds to spawn the instance on the hypervisor. [ 867.821682] env[61936]: DEBUG nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 867.822448] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881ba638-f2a9-434e-84bb-9b647a531747 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.844324] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252974, 'name': ReconfigVM_Task, 'duration_secs': 0.285149} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.844586] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 258fed0b-ba19-4c90-981a-96bb36ba8bc8/258fed0b-ba19-4c90-981a-96bb36ba8bc8.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.845748] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04b65a40-8fb0-4053-804d-b532aa8331b0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.852596] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 867.852596] env[61936]: value = "task-1252975" [ 867.852596] env[61936]: _type = "Task" [ 867.852596] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.862379] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252975, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.922659] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 868.008722] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.145200] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252973, 'name': CreateVM_Task, 'duration_secs': 0.936391} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.147928] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.148826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.149061] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.149428] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.149715] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eabbe3cc-a4e8-4055-a5b3-1de035e024d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.154342] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 868.154342] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52151e8c-8970-905f-3544-a602ae4d454f" [ 868.154342] env[61936]: _type = "Task" [ 868.154342] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.164958] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52151e8c-8970-905f-3544-a602ae4d454f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.181545] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526115c2-7c4b-1a81-12c9-e6b4aafd6502, 'name': SearchDatastore_Task, 'duration_secs': 0.056236} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.183806] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.184118] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.185511] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd6c0e66-a9dd-4163-a541-bce77b8e6fc4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.191805] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 868.191805] env[61936]: value = "task-1252976" [ 868.191805] env[61936]: _type = "Task" [ 868.191805] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.210501] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252976, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.247359] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1288f26a-d7b7-44af-8e71-eadce53eb5c9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.262201] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a0505a-46e2-4715-be6a-1e601b772c0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.276563] env[61936]: DEBUG nova.compute.manager [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received event network-changed-cf56feac-af85-4393-84a6-d364545347cc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 868.276776] env[61936]: DEBUG nova.compute.manager [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Refreshing instance network info cache due to event network-changed-cf56feac-af85-4393-84a6-d364545347cc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 868.276978] env[61936]: DEBUG oslo_concurrency.lockutils [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] Acquiring lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.312997] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e68fc5-cc0c-4eae-b2b2-8f33b1a94d7a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.321960] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8786e16-09a0-4660-ab8c-19cd42ca87f0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.344540] env[61936]: DEBUG nova.compute.provider_tree [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.348926] env[61936]: INFO nova.compute.manager [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Took 38.40 seconds to build instance. [ 868.362965] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252975, 'name': Rename_Task, 'duration_secs': 0.176766} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.363875] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.364408] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2275cca1-a560-4fa1-a906-f2ca06c612f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.374455] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 868.374455] env[61936]: value = "task-1252978" [ 868.374455] env[61936]: _type = "Task" [ 868.374455] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.385128] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252978, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.402140] env[61936]: DEBUG nova.network.neutron [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updated VIF entry in instance network info cache for port 12397fd2-b04a-4019-b4fb-ca0f7b7fc152. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.402445] env[61936]: DEBUG nova.network.neutron [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.666609] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52151e8c-8970-905f-3544-a602ae4d454f, 'name': SearchDatastore_Task, 'duration_secs': 0.010719} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.666979] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.667241] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.667535] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.668167] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.668167] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.668344] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebf934af-c95d-4be5-9402-5f139f8e0015 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.677358] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.677616] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.678479] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b2033b5-e38b-48f2-bd6b-be517bc2903b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.684166] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 868.684166] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52827a5e-74aa-f76e-35b9-d8e3bddc653a" [ 868.684166] env[61936]: _type = "Task" [ 868.684166] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.696529] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52827a5e-74aa-f76e-35b9-d8e3bddc653a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.705462] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252976, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481542} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.705769] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.706023] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.706299] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ceac5ced-1ea6-4a35-8169-096da9e44440 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.714738] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 868.714738] env[61936]: value = "task-1252979" [ 868.714738] env[61936]: _type = "Task" [ 868.714738] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.724778] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252979, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.745705] env[61936]: DEBUG nova.network.neutron [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Updating instance_info_cache with network_info: [{"id": "180a6391-4494-456c-b805-ec2ad1617c73", "address": "fa:16:3e:11:e3:2f", "network": {"id": "176b0129-7dfe-42db-a090-6360bc0134b3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1295754092", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f03f2cf1d4084fdca2664ccaac0a7671", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98011432-48cc-4ffd-a5a8-b96d2ea4424a", "external-id": "nsx-vlan-transportzone-745", "segmentation_id": 745, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap180a6391-44", "ovs_interfaceid": "180a6391-4494-456c-b805-ec2ad1617c73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf56feac-af85-4393-84a6-d364545347cc", "address": "fa:16:3e:a4:e0:31", "network": {"id": "757b0ece-6c31-4a8d-bf76-6cd285bbff68", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1064145213", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.246", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f03f2cf1d4084fdca2664ccaac0a7671", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf56feac-af", "ovs_interfaceid": "cf56feac-af85-4393-84a6-d364545347cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.850501] env[61936]: DEBUG nova.scheduler.client.report [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 868.854755] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f2897f03-b71f-4633-bc94-b8f24506b706 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.952s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.884588] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252978, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.905409] env[61936]: DEBUG oslo_concurrency.lockutils [req-b47cacfc-c862-425a-b7aa-07f4773797f5 req-08e16208-c73c-48c8-884e-5ef56c314c5a service nova] Releasing lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.932212] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 868.958839] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 868.959141] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.959297] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 868.959518] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.959674] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 868.959820] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 868.960569] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 868.960746] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 868.960917] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 868.961096] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 868.961271] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 868.962521] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f326b3f-a082-4c00-be1a-d38975d49a67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.972217] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0293abff-cca3-473f-aeb1-2c4ff9519429 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.194546] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52827a5e-74aa-f76e-35b9-d8e3bddc653a, 'name': SearchDatastore_Task, 'duration_secs': 0.010455} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.195358] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c71254c9-3f7c-42f3-b957-9dd7e077771a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.200974] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 869.200974] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f7d64e-d93c-0288-f94c-f4f0882ace06" [ 869.200974] env[61936]: _type = "Task" [ 869.200974] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.209298] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f7d64e-d93c-0288-f94c-f4f0882ace06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.228790] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252979, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088017} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.229144] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.229937] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f317854-07d8-4ca0-bf2b-9ba6f27e20a0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.251150] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.251644] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Releasing lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.251977] env[61936]: DEBUG nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Instance network_info: |[{"id": "180a6391-4494-456c-b805-ec2ad1617c73", "address": "fa:16:3e:11:e3:2f", "network": {"id": "176b0129-7dfe-42db-a090-6360bc0134b3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1295754092", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f03f2cf1d4084fdca2664ccaac0a7671", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98011432-48cc-4ffd-a5a8-b96d2ea4424a", "external-id": "nsx-vlan-transportzone-745", "segmentation_id": 745, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap180a6391-44", "ovs_interfaceid": "180a6391-4494-456c-b805-ec2ad1617c73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf56feac-af85-4393-84a6-d364545347cc", "address": "fa:16:3e:a4:e0:31", "network": {"id": "757b0ece-6c31-4a8d-bf76-6cd285bbff68", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1064145213", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.246", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f03f2cf1d4084fdca2664ccaac0a7671", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf56feac-af", "ovs_interfaceid": "cf56feac-af85-4393-84a6-d364545347cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 869.252226] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df0a1c0e-8b2f-4745-963c-ac12f4bc743a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.266363] env[61936]: DEBUG oslo_concurrency.lockutils [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] Acquired lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.266560] env[61936]: DEBUG nova.network.neutron [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Refreshing network info cache for port cf56feac-af85-4393-84a6-d364545347cc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 869.267689] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:e3:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98011432-48cc-4ffd-a5a8-b96d2ea4424a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '180a6391-4494-456c-b805-ec2ad1617c73', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:e0:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf56feac-af85-4393-84a6-d364545347cc', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.276758] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Creating folder: Project (f03f2cf1d4084fdca2664ccaac0a7671). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.280159] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5447bbca-b250-449a-8ae7-d77120d5343c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.290195] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 869.290195] env[61936]: value = "task-1252980" [ 869.290195] env[61936]: _type = "Task" [ 869.290195] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.293492] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Created folder: Project (f03f2cf1d4084fdca2664ccaac0a7671) in parent group-v269874. [ 869.293680] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Creating folder: Instances. Parent ref: group-v269953. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.294315] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-563c114e-1648-412c-bf3f-be3b478cde33 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.301792] env[61936]: DEBUG nova.compute.manager [req-510df508-3a04-4196-b91e-c0eede34921f req-58fd538a-0212-449c-88eb-63bb8e5dd8ff service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Received event network-vif-plugged-dc5d4628-264f-4f4e-9006-834caeb7b139 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 869.302077] env[61936]: DEBUG oslo_concurrency.lockutils [req-510df508-3a04-4196-b91e-c0eede34921f req-58fd538a-0212-449c-88eb-63bb8e5dd8ff service nova] Acquiring lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.305098] env[61936]: DEBUG oslo_concurrency.lockutils [req-510df508-3a04-4196-b91e-c0eede34921f req-58fd538a-0212-449c-88eb-63bb8e5dd8ff service nova] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.305098] env[61936]: DEBUG oslo_concurrency.lockutils [req-510df508-3a04-4196-b91e-c0eede34921f req-58fd538a-0212-449c-88eb-63bb8e5dd8ff service nova] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.305098] env[61936]: DEBUG nova.compute.manager [req-510df508-3a04-4196-b91e-c0eede34921f req-58fd538a-0212-449c-88eb-63bb8e5dd8ff service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] No waiting events found dispatching network-vif-plugged-dc5d4628-264f-4f4e-9006-834caeb7b139 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 869.305098] env[61936]: WARNING nova.compute.manager [req-510df508-3a04-4196-b91e-c0eede34921f req-58fd538a-0212-449c-88eb-63bb8e5dd8ff service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Received unexpected event network-vif-plugged-dc5d4628-264f-4f4e-9006-834caeb7b139 for instance with vm_state building and task_state spawning. [ 869.307987] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252980, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.318018] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Created folder: Instances in parent group-v269953. [ 869.318018] env[61936]: DEBUG oslo.service.loopingcall [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.318347] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.318823] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e97ee402-d0a2-4287-ad04-949ca9be79c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.342028] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.342028] env[61936]: value = "task-1252983" [ 869.342028] env[61936]: _type = "Task" [ 869.342028] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.350501] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252983, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.355821] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.356382] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 869.361108] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.849s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.361332] env[61936]: DEBUG nova.objects.instance [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lazy-loading 'resources' on Instance uuid 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.362422] env[61936]: DEBUG nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 869.388207] env[61936]: DEBUG oslo_vmware.api [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252978, 'name': PowerOnVM_Task, 'duration_secs': 0.636043} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.388482] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.388678] env[61936]: INFO nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Took 5.38 seconds to spawn the instance on the hypervisor. [ 869.388851] env[61936]: DEBUG nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 869.389674] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab4fc1b-559e-4f53-a4c0-8c86b82445ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.609498] env[61936]: DEBUG nova.network.neutron [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Updated VIF entry in instance network info cache for port cf56feac-af85-4393-84a6-d364545347cc. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.609927] env[61936]: DEBUG nova.network.neutron [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Updating instance_info_cache with network_info: [{"id": "180a6391-4494-456c-b805-ec2ad1617c73", "address": "fa:16:3e:11:e3:2f", "network": {"id": "176b0129-7dfe-42db-a090-6360bc0134b3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1295754092", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f03f2cf1d4084fdca2664ccaac0a7671", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98011432-48cc-4ffd-a5a8-b96d2ea4424a", "external-id": "nsx-vlan-transportzone-745", "segmentation_id": 745, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap180a6391-44", "ovs_interfaceid": "180a6391-4494-456c-b805-ec2ad1617c73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf56feac-af85-4393-84a6-d364545347cc", "address": "fa:16:3e:a4:e0:31", "network": {"id": "757b0ece-6c31-4a8d-bf76-6cd285bbff68", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1064145213", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.246", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "f03f2cf1d4084fdca2664ccaac0a7671", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf56feac-af", "ovs_interfaceid": "cf56feac-af85-4393-84a6-d364545347cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.712157] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f7d64e-d93c-0288-f94c-f4f0882ace06, 'name': SearchDatastore_Task, 'duration_secs': 0.035486} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.712428] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.712793] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 869.713183] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0050e8b0-9337-422c-bc09-84dc7563eab6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.721059] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 869.721059] env[61936]: value = "task-1252984" [ 869.721059] env[61936]: _type = "Task" [ 869.721059] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.730444] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.770844] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Successfully updated port: dc5d4628-264f-4f4e-9006-834caeb7b139 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.804118] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.853818] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252983, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.869611] env[61936]: DEBUG nova.compute.utils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.874497] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 869.874617] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.883129] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc6dbeb-88f7-495f-ac07-9478fe52a189 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.894546] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Suspending the VM {{(pid=61936) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 869.895584] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.895858] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-99b9d9bd-d196-418d-8818-917e84d626ba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.907924] env[61936]: INFO nova.compute.manager [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Took 21.89 seconds to build instance. [ 869.910524] env[61936]: DEBUG oslo_vmware.api [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 869.910524] env[61936]: value = "task-1252985" [ 869.910524] env[61936]: _type = "Task" [ 869.910524] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.927069] env[61936]: DEBUG oslo_vmware.api [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252985, 'name': SuspendVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.932280] env[61936]: DEBUG nova.policy [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6aec9537a6bc46b6b55b3163f1b59cc5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91a49a1ba2c34c8080cef82a7355cd95', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.113346] env[61936]: DEBUG oslo_concurrency.lockutils [req-f3788f08-e4bb-4436-9781-0aa674d79c5b req-989dbb55-4ea5-447f-af82-ded0dce07901 service nova] Releasing lock "refresh_cache-c4f17d38-25a3-4630-8bd4-bb40fa6fddee" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.235533] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252984, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.247838] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Successfully created port: 2a56df4a-9c49-4ffb-8c78-bdd4edd9714a {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.274059] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "refresh_cache-efdce04f-b0aa-49a6-94b8-1c960ac33fea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.274233] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "refresh_cache-efdce04f-b0aa-49a6-94b8-1c960ac33fea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.274391] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.304615] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252980, 'name': ReconfigVM_Task, 'duration_secs': 0.755131} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.307210] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Reconfigured VM instance instance-00000043 to attach disk [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.308834] env[61936]: DEBUG nova.compute.manager [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Received event network-changed-dc5d4628-264f-4f4e-9006-834caeb7b139 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 870.309025] env[61936]: DEBUG nova.compute.manager [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Refreshing instance network info cache due to event network-changed-dc5d4628-264f-4f4e-9006-834caeb7b139. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 870.309253] env[61936]: DEBUG oslo_concurrency.lockutils [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] Acquiring lock "refresh_cache-efdce04f-b0aa-49a6-94b8-1c960ac33fea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.310357] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79db3a4a-ea8f-40d3-aa6b-4251b36dee32 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.313184] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2a43ccc-6875-4f9c-afc1-4fe44b1b6751 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.323471] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc50c75-00e4-458d-9fde-bc14d3bc72d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.329384] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 870.329384] env[61936]: value = "task-1252986" [ 870.329384] env[61936]: _type = "Task" [ 870.329384] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.367238] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afa29b0-d20e-4041-a197-1e36729bdeb1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.370724] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252986, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.375635] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 870.387766] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252983, 'name': CreateVM_Task, 'duration_secs': 0.626035} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.389516] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbe8d6e-9898-418f-ae75-7fab82d738be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.395103] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.396571] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.396736] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.397219] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.397907] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af9c49d3-415a-481e-bc97-d46566005b9d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.411100] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e6904ce-54ea-4f90-8745-7bf8ed59bade tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.271s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.411624] env[61936]: DEBUG nova.compute.provider_tree [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.420349] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 870.420349] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526af74a-5da5-3ae1-c4fc-fb2861fa329f" [ 870.420349] env[61936]: _type = "Task" [ 870.420349] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.431018] env[61936]: DEBUG oslo_vmware.api [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252985, 'name': SuspendVM_Task} progress is 62%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.439521] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526af74a-5da5-3ae1-c4fc-fb2861fa329f, 'name': SearchDatastore_Task, 'duration_secs': 0.018449} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.439864] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.440169] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.440456] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.440614] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.440801] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.441100] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48aa2a0c-9760-417e-b1cb-d9caee582b3d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.454223] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.454223] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.454845] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-258b8c6f-e1d7-4302-9f21-35a59da31081 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.461484] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 870.461484] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd9770-8cdb-9b72-ed2f-3054caf4222c" [ 870.461484] env[61936]: _type = "Task" [ 870.461484] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.470497] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd9770-8cdb-9b72-ed2f-3054caf4222c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.735542] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252984, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.708693} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.735542] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.735856] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.735895] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51356608-1e7a-4059-b8eb-a9fc80c5f44a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.742650] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 870.742650] env[61936]: value = "task-1252988" [ 870.742650] env[61936]: _type = "Task" [ 870.742650] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.751995] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.816536] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.842581] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252986, 'name': Rename_Task, 'duration_secs': 0.360399} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.842901] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.843201] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2565c51b-daec-4c31-b8e1-12b6d9aaac17 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.850866] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 870.850866] env[61936]: value = "task-1252989" [ 870.850866] env[61936]: _type = "Task" [ 870.850866] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.862354] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252989, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.918599] env[61936]: DEBUG nova.scheduler.client.report [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 870.922465] env[61936]: DEBUG nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 870.933016] env[61936]: DEBUG oslo_vmware.api [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252985, 'name': SuspendVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.975148] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd9770-8cdb-9b72-ed2f-3054caf4222c, 'name': SearchDatastore_Task, 'duration_secs': 0.019192} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.976149] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92886635-47a1-451e-b90c-50fdf697d70f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.986052] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 870.986052] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d7b87e-f061-4ba0-1e88-73ab7b388236" [ 870.986052] env[61936]: _type = "Task" [ 870.986052] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.993552] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d7b87e-f061-4ba0-1e88-73ab7b388236, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.050856] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Updating instance_info_cache with network_info: [{"id": "dc5d4628-264f-4f4e-9006-834caeb7b139", "address": "fa:16:3e:bf:3f:36", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5d4628-26", "ovs_interfaceid": "dc5d4628-264f-4f4e-9006-834caeb7b139", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.254414] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060026} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.254729] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.255556] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb97bbde-68c4-4120-b619-d2a45da75afa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.280566] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.280876] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccef90d4-fe98-42d2-8c4b-0dc8cf21ce49 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.302764] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 871.302764] env[61936]: value = "task-1252990" [ 871.302764] env[61936]: _type = "Task" [ 871.302764] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.311269] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252990, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.361236] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252989, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.396257] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 871.421741] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 871.422022] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.422252] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 871.422476] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.422628] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 871.422776] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 871.422981] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 871.423176] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 871.423350] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 871.423560] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 871.423881] env[61936]: DEBUG nova.virt.hardware [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 871.424693] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27749136-c759-4c7f-825b-f9d51ce72902 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.431500] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.070s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.436043] env[61936]: DEBUG oslo_vmware.api [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1252985, 'name': SuspendVM_Task, 'duration_secs': 1.060316} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.437033] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.697s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.437033] env[61936]: DEBUG nova.objects.instance [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lazy-loading 'resources' on Instance uuid bdea3742-5789-4e97-b139-33cfee134843 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.438475] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Suspended the VM {{(pid=61936) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 871.438730] env[61936]: DEBUG nova.compute.manager [None req-633f985e-c00b-471a-a092-728d470cbc99 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 871.442031] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1ca89c-ff72-429a-93a0-a8c33e747636 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.449793] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f66841-d3bd-489d-af61-4035e51b1896 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.459194] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.460763] env[61936]: INFO nova.scheduler.client.report [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Deleted allocations for instance 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d [ 871.496419] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d7b87e-f061-4ba0-1e88-73ab7b388236, 'name': SearchDatastore_Task, 'duration_secs': 0.070026} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.496732] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.497061] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] c4f17d38-25a3-4630-8bd4-bb40fa6fddee/c4f17d38-25a3-4630-8bd4-bb40fa6fddee.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.497356] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8a9d95a-41f9-4bad-8628-ac0c93567818 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.507225] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 871.507225] env[61936]: value = "task-1252991" [ 871.507225] env[61936]: _type = "Task" [ 871.507225] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.517202] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.553415] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "refresh_cache-efdce04f-b0aa-49a6-94b8-1c960ac33fea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.553848] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Instance network_info: |[{"id": "dc5d4628-264f-4f4e-9006-834caeb7b139", "address": "fa:16:3e:bf:3f:36", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5d4628-26", "ovs_interfaceid": "dc5d4628-264f-4f4e-9006-834caeb7b139", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 871.554223] env[61936]: DEBUG oslo_concurrency.lockutils [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] Acquired lock "refresh_cache-efdce04f-b0aa-49a6-94b8-1c960ac33fea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.554431] env[61936]: DEBUG nova.network.neutron [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Refreshing network info cache for port dc5d4628-264f-4f4e-9006-834caeb7b139 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.555800] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:3f:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73e099e8-2acc-4628-a60d-0b4afa46b39d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc5d4628-264f-4f4e-9006-834caeb7b139', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 871.564457] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Creating folder: Project (91a49a1ba2c34c8080cef82a7355cd95). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 871.565680] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba59ab61-9363-4d58-b9b3-e4a561fad8a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.579826] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Created folder: Project (91a49a1ba2c34c8080cef82a7355cd95) in parent group-v269874. [ 871.580101] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Creating folder: Instances. Parent ref: group-v269956. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 871.580537] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-110f7b93-cc92-4340-be73-19949822ddab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.592831] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Created folder: Instances in parent group-v269956. [ 871.593102] env[61936]: DEBUG oslo.service.loopingcall [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.593439] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 871.593551] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95613695-f81d-4525-9236-d475a7299853 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.619436] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 871.619436] env[61936]: value = "task-1252994" [ 871.619436] env[61936]: _type = "Task" [ 871.619436] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.629194] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252994, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.816595] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252990, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.866987] env[61936]: DEBUG oslo_vmware.api [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1252989, 'name': PowerOnVM_Task, 'duration_secs': 0.892493} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.868498] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Successfully updated port: 2a56df4a-9c49-4ffb-8c78-bdd4edd9714a {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.870394] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.870702] env[61936]: INFO nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Took 5.40 seconds to spawn the instance on the hypervisor. [ 871.871284] env[61936]: DEBUG nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 871.872314] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc7e87e-c906-403a-9e35-a16aef6e1b30 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.982856] env[61936]: DEBUG oslo_concurrency.lockutils [None req-037f9f23-88f3-4a7f-ae51-9961b0de6498 tempest-ServersAaction247Test-1519022197 tempest-ServersAaction247Test-1519022197-project-member] Lock "644ccb42-44da-43e8-8b3b-9bfd2ac19a5d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.768s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.020563] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252991, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.130597] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1252994, 'name': CreateVM_Task, 'duration_secs': 0.444988} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.130765] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.131483] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.132154] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.132154] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.133077] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d6af40b-95e4-4a2e-b81c-69c95b81b616 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.139176] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 872.139176] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c6d9be-2626-cbc4-2bfa-359c0c2264f4" [ 872.139176] env[61936]: _type = "Task" [ 872.139176] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.153800] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c6d9be-2626-cbc4-2bfa-359c0c2264f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.293436] env[61936]: DEBUG nova.network.neutron [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Updated VIF entry in instance network info cache for port dc5d4628-264f-4f4e-9006-834caeb7b139. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.293834] env[61936]: DEBUG nova.network.neutron [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Updating instance_info_cache with network_info: [{"id": "dc5d4628-264f-4f4e-9006-834caeb7b139", "address": "fa:16:3e:bf:3f:36", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5d4628-26", "ovs_interfaceid": "dc5d4628-264f-4f4e-9006-834caeb7b139", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.309969] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4acc1389-90ce-4e52-8580-761221c02ba7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.320183] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252990, 'name': ReconfigVM_Task, 'duration_secs': 0.632516} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.322105] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Reconfigured VM instance instance-0000003e to attach disk [datastore2] e800f5e9-deb0-4592-aded-3768d238e311/e800f5e9-deb0-4592-aded-3768d238e311.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.322706] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a39324e0-391a-4ad7-b905-9d3528fcd4ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.325281] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5dceebc-2ede-4526-a571-ddf9ff3218d9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.364707] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82495799-4268-4d3e-8bc0-164d3ad1d4ba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.367706] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 872.367706] env[61936]: value = "task-1252995" [ 872.367706] env[61936]: _type = "Task" [ 872.367706] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.375996] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "refresh_cache-3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.375996] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "refresh_cache-3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.375996] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.380683] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7217108d-51d4-443b-b1bb-cc6725607e56 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.398807] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252995, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.402839] env[61936]: INFO nova.compute.manager [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Took 23.30 seconds to build instance. [ 872.410695] env[61936]: DEBUG nova.compute.provider_tree [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.496678] env[61936]: DEBUG nova.compute.manager [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Received event network-vif-plugged-2a56df4a-9c49-4ffb-8c78-bdd4edd9714a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 872.496761] env[61936]: DEBUG oslo_concurrency.lockutils [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] Acquiring lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.497053] env[61936]: DEBUG oslo_concurrency.lockutils [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.497293] env[61936]: DEBUG oslo_concurrency.lockutils [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.497524] env[61936]: DEBUG nova.compute.manager [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] No waiting events found dispatching network-vif-plugged-2a56df4a-9c49-4ffb-8c78-bdd4edd9714a {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 872.497764] env[61936]: WARNING nova.compute.manager [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Received unexpected event network-vif-plugged-2a56df4a-9c49-4ffb-8c78-bdd4edd9714a for instance with vm_state building and task_state spawning. [ 872.497874] env[61936]: DEBUG nova.compute.manager [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Received event network-changed-2a56df4a-9c49-4ffb-8c78-bdd4edd9714a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 872.498010] env[61936]: DEBUG nova.compute.manager [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Refreshing instance network info cache due to event network-changed-2a56df4a-9c49-4ffb-8c78-bdd4edd9714a. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 872.498246] env[61936]: DEBUG oslo_concurrency.lockutils [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] Acquiring lock "refresh_cache-3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.524432] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252991, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523069} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.524699] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] c4f17d38-25a3-4630-8bd4-bb40fa6fddee/c4f17d38-25a3-4630-8bd4-bb40fa6fddee.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.524909] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.525182] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74cf762d-53f3-4a38-89e1-892f6190d168 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.533362] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 872.533362] env[61936]: value = "task-1252996" [ 872.533362] env[61936]: _type = "Task" [ 872.533362] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.543012] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252996, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.651459] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c6d9be-2626-cbc4-2bfa-359c0c2264f4, 'name': SearchDatastore_Task, 'duration_secs': 0.024671} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.652587] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.652587] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.652587] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.652587] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.652587] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.652861] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a1402b6-4303-4696-b9b6-9d3f98f3daf3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.669184] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.669443] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.670560] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d52c7bb-5ec3-4ed9-8c28-388f672eb5f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.677214] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 872.677214] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cbbd52-c32d-8a15-ddf1-5845cf1af4ca" [ 872.677214] env[61936]: _type = "Task" [ 872.677214] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.691272] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cbbd52-c32d-8a15-ddf1-5845cf1af4ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.800671] env[61936]: DEBUG oslo_concurrency.lockutils [req-909f4fd0-a87b-41ce-974d-cf6ef3d2e9da req-4a6f162e-fd10-44f9-be4f-0569fd311b8c service nova] Releasing lock "refresh_cache-efdce04f-b0aa-49a6-94b8-1c960ac33fea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.880229] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252995, 'name': Rename_Task, 'duration_secs': 0.167691} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.880515] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.881068] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6d76555-5b47-4016-aa6f-96c1f7b657e6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.888028] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 872.888028] env[61936]: value = "task-1252997" [ 872.888028] env[61936]: _type = "Task" [ 872.888028] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.896483] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.902068] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b614bb8b-a371-4b31-9686-dcaa6455fcb9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "aa37292e-8499-46aa-8021-cb8d88b5a35a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.503s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.916873] env[61936]: DEBUG nova.scheduler.client.report [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 872.939164] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.043662] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252996, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071573} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.043964] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.044761] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c11f056-a109-4252-8205-fb5e575f999b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.075623] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] c4f17d38-25a3-4630-8bd4-bb40fa6fddee/c4f17d38-25a3-4630-8bd4-bb40fa6fddee.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.078314] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19a10ee2-d680-4f11-9501-8bb1a0837140 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.099886] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 873.099886] env[61936]: value = "task-1252998" [ 873.099886] env[61936]: _type = "Task" [ 873.099886] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.111802] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252998, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.146504] env[61936]: DEBUG nova.network.neutron [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Updating instance_info_cache with network_info: [{"id": "2a56df4a-9c49-4ffb-8c78-bdd4edd9714a", "address": "fa:16:3e:7a:69:5e", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a56df4a-9c", "ovs_interfaceid": "2a56df4a-9c49-4ffb-8c78-bdd4edd9714a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.171259] env[61936]: INFO nova.compute.manager [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Rebuilding instance [ 873.189170] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cbbd52-c32d-8a15-ddf1-5845cf1af4ca, 'name': SearchDatastore_Task, 'duration_secs': 0.016323} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.191996] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8538f73b-8e76-4c6c-989f-dd734d0fc97c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.197644] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 873.197644] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c5fe0a-82f4-d5b5-6903-5f572cee53c1" [ 873.197644] env[61936]: _type = "Task" [ 873.197644] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.210086] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c5fe0a-82f4-d5b5-6903-5f572cee53c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.222336] env[61936]: DEBUG nova.compute.manager [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.223180] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2348b45e-680c-4f67-842a-b9e00ee0d94d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.271544] env[61936]: DEBUG nova.compute.manager [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.272682] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5658bc-ce13-480d-b8e4-c3dbd0192841 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.402191] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252997, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.404762] env[61936]: DEBUG nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 873.422604] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.986s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.425377] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.142s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.425643] env[61936]: DEBUG nova.objects.instance [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lazy-loading 'resources' on Instance uuid 301b0504-5cf9-44e0-bd3e-342f8a89278b {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.446538] env[61936]: INFO nova.scheduler.client.report [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleted allocations for instance bdea3742-5789-4e97-b139-33cfee134843 [ 873.611274] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252998, 'name': ReconfigVM_Task, 'duration_secs': 0.413371} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.611728] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Reconfigured VM instance instance-00000041 to attach disk [datastore2] c4f17d38-25a3-4630-8bd4-bb40fa6fddee/c4f17d38-25a3-4630-8bd4-bb40fa6fddee.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.612478] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83516f95-6689-4968-b51d-60250d5de8fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.619852] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 873.619852] env[61936]: value = "task-1252999" [ 873.619852] env[61936]: _type = "Task" [ 873.619852] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.628873] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252999, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.649774] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "refresh_cache-3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.650164] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Instance network_info: |[{"id": "2a56df4a-9c49-4ffb-8c78-bdd4edd9714a", "address": "fa:16:3e:7a:69:5e", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a56df4a-9c", "ovs_interfaceid": "2a56df4a-9c49-4ffb-8c78-bdd4edd9714a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 873.650513] env[61936]: DEBUG oslo_concurrency.lockutils [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] Acquired lock "refresh_cache-3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.650688] env[61936]: DEBUG nova.network.neutron [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Refreshing network info cache for port 2a56df4a-9c49-4ffb-8c78-bdd4edd9714a {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.652124] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:69:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73e099e8-2acc-4628-a60d-0b4afa46b39d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a56df4a-9c49-4ffb-8c78-bdd4edd9714a', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.660227] env[61936]: DEBUG oslo.service.loopingcall [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.663686] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.664241] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7e6aed6-30e7-405d-a0f8-f75d39c1c852 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.686232] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.686232] env[61936]: value = "task-1253000" [ 873.686232] env[61936]: _type = "Task" [ 873.686232] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.694408] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253000, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.713196] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c5fe0a-82f4-d5b5-6903-5f572cee53c1, 'name': SearchDatastore_Task, 'duration_secs': 0.014986} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.713511] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.713801] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] efdce04f-b0aa-49a6-94b8-1c960ac33fea/efdce04f-b0aa-49a6-94b8-1c960ac33fea.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.714097] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13d4981d-02f6-45f0-ac82-2ca5e0c13ed4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.726696] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 873.726696] env[61936]: value = "task-1253001" [ 873.726696] env[61936]: _type = "Task" [ 873.726696] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.742945] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253001, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.784954] env[61936]: INFO nova.compute.manager [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] instance snapshotting [ 873.785594] env[61936]: WARNING nova.compute.manager [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 873.789471] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f0c725-fc56-496b-b64d-f09af01161c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.820850] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812ffb27-b0f1-4d80-8c9a-84397b8c894e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.905763] env[61936]: DEBUG oslo_vmware.api [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1252997, 'name': PowerOnVM_Task, 'duration_secs': 0.561533} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.906078] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.906384] env[61936]: DEBUG nova.compute.manager [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.907191] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3719fe49-ce49-4dd5-bdbd-90ebf9b15c12 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.947535] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.959145] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0043b719-c0c3-4d27-ac3f-b34f13afdd0a tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "bdea3742-5789-4e97-b139-33cfee134843" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.714s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.089671] env[61936]: DEBUG nova.network.neutron [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Updated VIF entry in instance network info cache for port 2a56df4a-9c49-4ffb-8c78-bdd4edd9714a. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.094150] env[61936]: DEBUG nova.network.neutron [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Updating instance_info_cache with network_info: [{"id": "2a56df4a-9c49-4ffb-8c78-bdd4edd9714a", "address": "fa:16:3e:7a:69:5e", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a56df4a-9c", "ovs_interfaceid": "2a56df4a-9c49-4ffb-8c78-bdd4edd9714a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.142135] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1252999, 'name': Rename_Task, 'duration_secs': 0.252811} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.146435] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.147041] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4c09c72-2f5f-4356-bd21-1c72d58d08dc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.157301] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 874.157301] env[61936]: value = "task-1253002" [ 874.157301] env[61936]: _type = "Task" [ 874.157301] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.171777] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1253002, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.202072] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253000, 'name': CreateVM_Task, 'duration_secs': 0.418601} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.207090] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 874.207090] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.207090] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.207090] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 874.207090] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-727ebf24-9274-4f04-8476-90aff7639ad2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.214202] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 874.214202] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5212cc31-6c6f-923d-4f88-e43e8e7da79b" [ 874.214202] env[61936]: _type = "Task" [ 874.214202] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.229634] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5212cc31-6c6f-923d-4f88-e43e8e7da79b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.239025] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.239025] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8bf9483-5ff4-403d-95b7-1a2a18cf663b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.244080] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253001, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.246726] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 874.246726] env[61936]: value = "task-1253003" [ 874.246726] env[61936]: _type = "Task" [ 874.246726] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.258437] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.336042] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Creating Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 874.336042] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-59c15152-2813-482d-97f2-23c64286e373 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.347053] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 874.347053] env[61936]: value = "task-1253004" [ 874.347053] env[61936]: _type = "Task" [ 874.347053] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.359766] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253004, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.377050] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e64a49-ee9c-4db1-8c62-aa05156f806d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.391148] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8d6438-2b23-43b1-a4ba-bd59fb6f678f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.443226] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1086030a-2844-4ac7-a29a-bff456782a2e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.447127] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.454701] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5941686-e59a-43cb-8463-e59214cc9285 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.476730] env[61936]: DEBUG nova.compute.provider_tree [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.597967] env[61936]: DEBUG oslo_concurrency.lockutils [req-d76238c6-b542-4e77-b5a1-b332336add30 req-f6cd6706-0617-44b0-935c-600a32013faf service nova] Releasing lock "refresh_cache-3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.678558] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1253002, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.730114] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5212cc31-6c6f-923d-4f88-e43e8e7da79b, 'name': SearchDatastore_Task, 'duration_secs': 0.022423} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.731136] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.731487] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.731749] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.731894] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.732090] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.734040] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79989468-0d04-4c48-9760-4672d0c09f02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.746855] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253001, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526487} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.747899] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] efdce04f-b0aa-49a6-94b8-1c960ac33fea/efdce04f-b0aa-49a6-94b8-1c960ac33fea.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.747899] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.752461] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75e5fb4f-9b64-419f-941a-46ec9af0b0eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.757374] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.757374] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.757374] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dac9c3cd-c08f-4577-b2f2-2a5f339eee20 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.767954] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253003, 'name': PowerOffVM_Task, 'duration_secs': 0.192018} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.772499] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.772499] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.772968] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 874.772968] env[61936]: value = "task-1253005" [ 874.772968] env[61936]: _type = "Task" [ 874.772968] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.773451] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 874.773451] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52288739-2ce9-0b41-b43a-0f1455f3a0e6" [ 874.773451] env[61936]: _type = "Task" [ 874.773451] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.774764] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6beb6d-ca21-4826-a460-003241f85b2e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.797141] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253005, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.797625] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.801428] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b68c4b2f-b200-4695-b9da-528a88c885cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.804774] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52288739-2ce9-0b41-b43a-0f1455f3a0e6, 'name': SearchDatastore_Task, 'duration_secs': 0.020834} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.806266] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc4e5977-2582-467f-90bc-51fc51d4e311 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.813987] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 874.813987] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52798dea-8188-bae4-c090-4a48f6bfd77f" [ 874.813987] env[61936]: _type = "Task" [ 874.813987] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.825498] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52798dea-8188-bae4-c090-4a48f6bfd77f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.844782] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.845420] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.845535] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleting the datastore file [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.845885] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36e0983e-75ca-44aa-b3f7-44fadcb1ae83 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.862459] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253004, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.866581] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 874.866581] env[61936]: value = "task-1253007" [ 874.866581] env[61936]: _type = "Task" [ 874.866581] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.876943] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.980513] env[61936]: DEBUG nova.scheduler.client.report [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 874.995812] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquiring lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.996222] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.172176] env[61936]: DEBUG oslo_vmware.api [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1253002, 'name': PowerOnVM_Task, 'duration_secs': 0.633373} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.172568] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 875.172836] env[61936]: INFO nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Took 13.60 seconds to spawn the instance on the hypervisor. [ 875.173090] env[61936]: DEBUG nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 875.173985] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4315b20d-2794-4a21-899d-dfe280d0416c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.179491] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "e800f5e9-deb0-4592-aded-3768d238e311" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.179727] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "e800f5e9-deb0-4592-aded-3768d238e311" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.179959] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "e800f5e9-deb0-4592-aded-3768d238e311-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.180125] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "e800f5e9-deb0-4592-aded-3768d238e311-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.180328] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "e800f5e9-deb0-4592-aded-3768d238e311-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.186814] env[61936]: INFO nova.compute.manager [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Terminating instance [ 875.295028] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253005, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106267} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.295028] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.295028] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58fb178-3913-4d86-b62d-75ee95608d24 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.319169] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] efdce04f-b0aa-49a6-94b8-1c960ac33fea/efdce04f-b0aa-49a6-94b8-1c960ac33fea.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.319169] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f29c2f4a-f5b1-4cfc-b804-c6f25278e6ba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.346412] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52798dea-8188-bae4-c090-4a48f6bfd77f, 'name': SearchDatastore_Task, 'duration_secs': 0.021765} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.347842] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.348422] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e/3b09b9f5-9084-4bf4-a441-fb4b2d13c73e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.348525] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 875.348525] env[61936]: value = "task-1253008" [ 875.348525] env[61936]: _type = "Task" [ 875.348525] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.348748] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24316359-b5d7-46f2-8fa5-5bbbe569a13f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.365692] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253008, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.372571] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253004, 'name': CreateSnapshot_Task, 'duration_secs': 0.678678} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.372571] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 875.372571] env[61936]: value = "task-1253009" [ 875.372571] env[61936]: _type = "Task" [ 875.372571] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.374378] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Created Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 875.375260] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1ec2e8-12ec-4b4d-904f-3b165bd9c71f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.391572] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280174} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.398541] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.398541] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.398541] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.400436] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.496368] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.071s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.499172] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.593s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.500866] env[61936]: INFO nova.compute.claims [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.533065] env[61936]: INFO nova.scheduler.client.report [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Deleted allocations for instance 301b0504-5cf9-44e0-bd3e-342f8a89278b [ 875.698802] env[61936]: DEBUG nova.compute.manager [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 875.698802] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.698802] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c75ee26-6d8a-4ccd-bac5-dccd63e64e92 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.704780] env[61936]: INFO nova.compute.manager [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Took 33.32 seconds to build instance. [ 875.715490] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 875.715490] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8451f6b-8e16-4258-80ed-c9d6ad634b0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.725018] env[61936]: DEBUG oslo_vmware.api [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 875.725018] env[61936]: value = "task-1253010" [ 875.725018] env[61936]: _type = "Task" [ 875.725018] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.736381] env[61936]: DEBUG oslo_vmware.api [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.865176] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253008, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.885691] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253009, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.892495] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.892495] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.915325] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Creating linked-clone VM from snapshot {{(pid=61936) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 875.922025] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-817a5686-4029-4c40-a025-a8a2eb007700 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.934911] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 875.934911] env[61936]: value = "task-1253011" [ 875.934911] env[61936]: _type = "Task" [ 875.934911] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.947564] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253011, 'name': CloneVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.043027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e85e3b5-c55c-4736-8608-a8d91b670c97 tempest-ServersV294TestFqdnHostnames-801876851 tempest-ServersV294TestFqdnHostnames-801876851-project-member] Lock "301b0504-5cf9-44e0-bd3e-342f8a89278b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.768s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.207594] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c7e1673a-c0f7-486d-a8a7-1a793c4d71da tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.060s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.235562] env[61936]: DEBUG oslo_vmware.api [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253010, 'name': PowerOffVM_Task, 'duration_secs': 0.376111} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.235820] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.235977] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.236280] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4ea860dc-112f-4f61-9eed-46cedbdba8e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.328328] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.328560] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.328738] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleting the datastore file [datastore2] e800f5e9-deb0-4592-aded-3768d238e311 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.329014] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4cbdfe78-5e84-485e-a9d5-e3007e6b142e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.335866] env[61936]: DEBUG oslo_vmware.api [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 876.335866] env[61936]: value = "task-1253013" [ 876.335866] env[61936]: _type = "Task" [ 876.335866] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.344642] env[61936]: DEBUG oslo_vmware.api [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.362641] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253008, 'name': ReconfigVM_Task, 'duration_secs': 0.724153} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.362923] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Reconfigured VM instance instance-00000044 to attach disk [datastore1] efdce04f-b0aa-49a6-94b8-1c960ac33fea/efdce04f-b0aa-49a6-94b8-1c960ac33fea.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.363567] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81cb8300-6a1f-4b18-9e5b-ec793727ae62 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.371488] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 876.371488] env[61936]: value = "task-1253014" [ 876.371488] env[61936]: _type = "Task" [ 876.371488] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.381984] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253014, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.388327] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733504} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.388611] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e/3b09b9f5-9084-4bf4-a441-fb4b2d13c73e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 876.388823] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.389093] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94fd6e9c-b234-45d0-8139-99879fb7c021 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.396347] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 876.396347] env[61936]: value = "task-1253015" [ 876.396347] env[61936]: _type = "Task" [ 876.396347] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.406127] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.451490] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253011, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.456382] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 876.457689] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.457689] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.457689] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.457689] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.457689] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 876.457689] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 876.457972] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 876.458202] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 876.458464] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 876.459036] env[61936]: DEBUG nova.virt.hardware [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 876.459695] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e4d1eb-442c-49f4-841c-f9a22a1f42b0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.471152] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478c6c2d-428d-4549-92a5-fc2ca8555d74 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.496834] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Instance VIF info [] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.504623] env[61936]: DEBUG oslo.service.loopingcall [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.506399] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.506754] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45ea3b98-3bd1-4aa5-845e-77d78e0a17b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.536597] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.536597] env[61936]: value = "task-1253016" [ 876.536597] env[61936]: _type = "Task" [ 876.536597] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.548492] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253016, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.585346] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "fa845f8e-957c-4c0b-a647-190c32989dcd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.585751] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.710274] env[61936]: DEBUG nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 876.848845] env[61936]: DEBUG oslo_vmware.api [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287537} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.852906] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.853118] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 876.853308] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 876.853481] env[61936]: INFO nova.compute.manager [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Took 1.16 seconds to destroy the instance on the hypervisor. [ 876.853731] env[61936]: DEBUG oslo.service.loopingcall [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.854863] env[61936]: DEBUG nova.compute.manager [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 876.854967] env[61936]: DEBUG nova.network.neutron [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 876.883077] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253014, 'name': Rename_Task, 'duration_secs': 0.227726} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.885874] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.887188] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3a257cf-5e1e-4d06-8196-0cf69d571f05 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.896015] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 876.896015] env[61936]: value = "task-1253017" [ 876.896015] env[61936]: _type = "Task" [ 876.896015] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.912921] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111573} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.915754] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.916019] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253017, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.917585] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69586854-9383-4c13-a54a-6017ed32c960 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.949768] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e/3b09b9f5-9084-4bf4-a441-fb4b2d13c73e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.957643] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7840beed-a76e-47b1-ba97-5f5678f75691 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.980323] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253011, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.982032] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 876.982032] env[61936]: value = "task-1253018" [ 876.982032] env[61936]: _type = "Task" [ 876.982032] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.988085] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.988397] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.988610] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.988801] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.988981] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.991874] env[61936]: INFO nova.compute.manager [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Terminating instance [ 876.998453] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253018, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.010612] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b539adb-3ffd-4a0b-858b-c1c735b88a48 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.021333] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1053b192-ffdb-4eb0-9f3b-296da37d0b29 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.061773] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e3352d-6a95-4b43-833f-684c11c0fa08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.073670] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c65cd3-e994-484c-810d-b35eb59394c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.078198] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253016, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.091366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "77a59650-d28b-4862-bb35-6910e6f7d6c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.091366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.091366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "77a59650-d28b-4862-bb35-6910e6f7d6c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.091366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.091366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.093154] env[61936]: DEBUG nova.compute.provider_tree [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.094598] env[61936]: INFO nova.compute.manager [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Terminating instance [ 877.244651] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.300439] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.300439] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.300607] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.300782] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.300949] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.303402] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.303697] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.304212] env[61936]: INFO nova.compute.manager [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Terminating instance [ 877.411207] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253017, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.448824] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253011, 'name': CloneVM_Task, 'duration_secs': 1.487111} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.449186] env[61936]: INFO nova.virt.vmwareapi.vmops [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Created linked-clone VM from snapshot [ 877.453155] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc2c4c7-ea6f-41e6-8fe9-2a4fb4d7cb4e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.458431] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Uploading image 258fa9b9-a92c-4360-8f49-b8b720244f4e {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 877.464252] env[61936]: DEBUG nova.compute.manager [req-b84f4b9c-460a-4b35-bd42-3458abee5d94 req-9a027e64-c797-4aea-ac1a-46a56747a25d service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Received event network-vif-deleted-48edeab7-9688-4106-9b4a-9e53a69aa280 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 877.464440] env[61936]: INFO nova.compute.manager [req-b84f4b9c-460a-4b35-bd42-3458abee5d94 req-9a027e64-c797-4aea-ac1a-46a56747a25d service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Neutron deleted interface 48edeab7-9688-4106-9b4a-9e53a69aa280; detaching it from the instance and deleting it from the info cache [ 877.464669] env[61936]: DEBUG nova.network.neutron [req-b84f4b9c-460a-4b35-bd42-3458abee5d94 req-9a027e64-c797-4aea-ac1a-46a56747a25d service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.489309] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 877.489309] env[61936]: value = "vm-269961" [ 877.489309] env[61936]: _type = "VirtualMachine" [ 877.489309] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 877.490144] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-485910ef-ad7e-4f3f-b808-695ebba1f03c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.498578] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253018, 'name': ReconfigVM_Task, 'duration_secs': 0.41378} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.500589] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e/3b09b9f5-9084-4bf4-a441-fb4b2d13c73e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.501474] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lease: (returnval){ [ 877.501474] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526dcf8a-5165-9238-b109-f9b71305ec2f" [ 877.501474] env[61936]: _type = "HttpNfcLease" [ 877.501474] env[61936]: } obtained for exporting VM: (result){ [ 877.501474] env[61936]: value = "vm-269961" [ 877.501474] env[61936]: _type = "VirtualMachine" [ 877.501474] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 877.501959] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the lease: (returnval){ [ 877.501959] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526dcf8a-5165-9238-b109-f9b71305ec2f" [ 877.501959] env[61936]: _type = "HttpNfcLease" [ 877.501959] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 877.502851] env[61936]: DEBUG nova.compute.manager [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 877.502851] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.503205] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-97aaf599-805c-4cfd-a8e0-c019342c8c1a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.505635] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd76dd1f-a591-4dfb-9fd5-6f58df5edd76 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.524070] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.524070] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526dcf8a-5165-9238-b109-f9b71305ec2f" [ 877.524070] env[61936]: _type = "HttpNfcLease" [ 877.524070] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 877.524070] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.524070] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 877.524070] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526dcf8a-5165-9238-b109-f9b71305ec2f" [ 877.524070] env[61936]: _type = "HttpNfcLease" [ 877.524070] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 877.524070] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c1ee5af-11e3-4913-86c7-39fde0bef7a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.525199] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 877.525199] env[61936]: value = "task-1253020" [ 877.525199] env[61936]: _type = "Task" [ 877.525199] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.525924] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fd0b1d-5261-4b6a-8404-9ce2ccbe9fc3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.536900] env[61936]: DEBUG oslo_vmware.api [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 877.536900] env[61936]: value = "task-1253021" [ 877.536900] env[61936]: _type = "Task" [ 877.536900] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.544992] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d4a9a-bfa3-e077-c5ee-528705949da0/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 877.545359] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d4a9a-bfa3-e077-c5ee-528705949da0/disk-0.vmdk for reading. {{(pid=61936) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 877.546839] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253020, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.614096] env[61936]: DEBUG nova.scheduler.client.report [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 877.618663] env[61936]: DEBUG nova.compute.manager [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 877.618906] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.629073] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8df8bd7-e025-4eff-86c2-41eb90179fcd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.632700] env[61936]: DEBUG oslo_vmware.api [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1253021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.641122] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253016, 'name': CreateVM_Task, 'duration_secs': 0.540244} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.643413] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.643748] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.644204] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.644377] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.644711] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.644935] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed6741b7-3a16-48e4-b795-f595b572db67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.646900] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4102990e-efdd-4f92-890a-e42e1051ebc3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.652413] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 877.652413] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bc36de-819e-1847-a1d6-f10921958645" [ 877.652413] env[61936]: _type = "Task" [ 877.652413] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.658170] env[61936]: DEBUG oslo_vmware.api [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 877.658170] env[61936]: value = "task-1253022" [ 877.658170] env[61936]: _type = "Task" [ 877.658170] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.665075] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bc36de-819e-1847-a1d6-f10921958645, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.670590] env[61936]: DEBUG oslo_vmware.api [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.688416] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c2fb42a9-baeb-4142-bef6-693ca841b20b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.808396] env[61936]: DEBUG nova.compute.manager [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 877.808633] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.809567] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a80b38-6b38-4235-a947-ec7b27a0dd9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.819125] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.819435] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8a0136d-bb0a-4578-b1c5-150158b85b54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.829053] env[61936]: DEBUG oslo_vmware.api [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 877.829053] env[61936]: value = "task-1253023" [ 877.829053] env[61936]: _type = "Task" [ 877.829053] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.840045] env[61936]: DEBUG oslo_vmware.api [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253023, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.895605] env[61936]: DEBUG nova.network.neutron [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.915960] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253017, 'name': PowerOnVM_Task, 'duration_secs': 0.754389} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.916969] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.917155] env[61936]: INFO nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Took 8.98 seconds to spawn the instance on the hypervisor. [ 877.917333] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 877.918329] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06196b12-bed8-48f5-8c34-0a6013b01921 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.969986] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4398d6e0-f123-4127-9553-90404004778b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.987820] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6663028-a137-4795-b781-12f3c77c69e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.035137] env[61936]: DEBUG nova.compute.manager [req-b84f4b9c-460a-4b35-bd42-3458abee5d94 req-9a027e64-c797-4aea-ac1a-46a56747a25d service nova] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Detach interface failed, port_id=48edeab7-9688-4106-9b4a-9e53a69aa280, reason: Instance e800f5e9-deb0-4592-aded-3768d238e311 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 878.045282] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253020, 'name': Rename_Task, 'duration_secs': 0.211478} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.049664] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.050408] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f26ad682-a9c7-4622-9439-f4ae11e2af8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.058693] env[61936]: DEBUG oslo_vmware.api [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1253021, 'name': PowerOffVM_Task, 'duration_secs': 0.293845} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.060616] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.061040] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.061978] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 878.061978] env[61936]: value = "task-1253024" [ 878.061978] env[61936]: _type = "Task" [ 878.061978] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.062234] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-354042ec-4a51-4799-a1a8-c7a0e31a38e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.075409] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253024, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.119799] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.120410] env[61936]: DEBUG nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 878.124161] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.321s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.125817] env[61936]: INFO nova.compute.claims [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.178476] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bc36de-819e-1847-a1d6-f10921958645, 'name': SearchDatastore_Task, 'duration_secs': 0.012127} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.183526] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.183947] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.185282] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.185282] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.185282] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.185552] env[61936]: DEBUG oslo_vmware.api [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253022, 'name': PowerOffVM_Task, 'duration_secs': 0.220977} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.185606] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc08ffc0-72fd-431c-a0f4-9b9c2da4e36d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.188126] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.188536] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.189307] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07457f41-e777-4105-a342-56adebd79480 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.193396] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 878.193699] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 878.193928] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Deleting the datastore file [datastore2] c4f17d38-25a3-4630-8bd4-bb40fa6fddee {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.195182] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d457cd37-9339-4603-830c-a9a315b573e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.203754] env[61936]: DEBUG oslo_vmware.api [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for the task: (returnval){ [ 878.203754] env[61936]: value = "task-1253027" [ 878.203754] env[61936]: _type = "Task" [ 878.203754] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.205205] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.205530] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.209523] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d98f7bc-7663-4398-bd71-a68c8bbe7b58 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.219271] env[61936]: DEBUG oslo_vmware.api [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1253027, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.222447] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 878.222447] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f30a62-1341-764f-dcb3-de690069997e" [ 878.222447] env[61936]: _type = "Task" [ 878.222447] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.234300] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f30a62-1341-764f-dcb3-de690069997e, 'name': SearchDatastore_Task, 'duration_secs': 0.009911} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.235355] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d5fe902-8ea8-4817-a964-b755299a15c5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.242115] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 878.242115] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b7a70b-7da1-c2e6-056d-962302c6c75c" [ 878.242115] env[61936]: _type = "Task" [ 878.242115] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.252852] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b7a70b-7da1-c2e6-056d-962302c6c75c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.285177] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 878.285403] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 878.285685] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleting the datastore file [datastore1] 77a59650-d28b-4862-bb35-6910e6f7d6c9 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.285929] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7998128f-50f9-4893-9b3c-81ebfb1a4a61 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.294559] env[61936]: DEBUG oslo_vmware.api [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 878.294559] env[61936]: value = "task-1253028" [ 878.294559] env[61936]: _type = "Task" [ 878.294559] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.306663] env[61936]: DEBUG oslo_vmware.api [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253028, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.341114] env[61936]: DEBUG oslo_vmware.api [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253023, 'name': PowerOffVM_Task, 'duration_secs': 0.243884} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.341114] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.341114] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.341114] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdd0560c-fa82-4e26-af4c-b32ef2ca1d83 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.399539] env[61936]: INFO nova.compute.manager [-] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Took 1.54 seconds to deallocate network for instance. [ 878.415258] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 878.415751] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 878.416563] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleting the datastore file [datastore2] fafc995e-4c93-4ca0-b078-24a1ae0ab427 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 878.417226] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-672c1171-adba-4386-8dad-42b599121dc6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.426769] env[61936]: DEBUG oslo_vmware.api [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for the task: (returnval){ [ 878.426769] env[61936]: value = "task-1253030" [ 878.426769] env[61936]: _type = "Task" [ 878.426769] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.448968] env[61936]: DEBUG oslo_vmware.api [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.455284] env[61936]: INFO nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Took 27.50 seconds to build instance. [ 878.576232] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253024, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.626263] env[61936]: DEBUG nova.compute.utils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.629248] env[61936]: DEBUG nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 878.629248] env[61936]: DEBUG nova.network.neutron [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 878.711945] env[61936]: DEBUG nova.policy [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd49bc89c58f24393952fd61c649e42cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3202ab76c58f4615bf87a324602770af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 878.721137] env[61936]: DEBUG oslo_vmware.api [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Task: {'id': task-1253027, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220125} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.721782] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.722198] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.722542] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.723243] env[61936]: INFO nova.compute.manager [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Took 1.22 seconds to destroy the instance on the hypervisor. [ 878.723709] env[61936]: DEBUG oslo.service.loopingcall [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.723983] env[61936]: DEBUG nova.compute.manager [-] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 878.724141] env[61936]: DEBUG nova.network.neutron [-] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.754579] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b7a70b-7da1-c2e6-056d-962302c6c75c, 'name': SearchDatastore_Task, 'duration_secs': 0.028509} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.754974] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.755413] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.755768] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f57a64d1-8903-4df6-98fc-edd52bd3a270 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.764607] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 878.764607] env[61936]: value = "task-1253031" [ 878.764607] env[61936]: _type = "Task" [ 878.764607] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.777446] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253031, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.807218] env[61936]: DEBUG oslo_vmware.api [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253028, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154948} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.809743] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.810182] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.810259] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.810451] env[61936]: INFO nova.compute.manager [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Took 1.19 seconds to destroy the instance on the hypervisor. [ 878.810862] env[61936]: DEBUG oslo.service.loopingcall [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.811158] env[61936]: DEBUG nova.compute.manager [-] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 878.811273] env[61936]: DEBUG nova.network.neutron [-] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.907902] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.939087] env[61936]: DEBUG oslo_vmware.api [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Task: {'id': task-1253030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148769} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.939791] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.941498] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.941498] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.941498] env[61936]: INFO nova.compute.manager [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Took 1.13 seconds to destroy the instance on the hypervisor. [ 878.941498] env[61936]: DEBUG oslo.service.loopingcall [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.941498] env[61936]: DEBUG nova.compute.manager [-] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 878.941498] env[61936]: DEBUG nova.network.neutron [-] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.960625] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.027s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.077560] env[61936]: DEBUG oslo_vmware.api [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253024, 'name': PowerOnVM_Task, 'duration_secs': 0.587909} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.077560] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.077560] env[61936]: INFO nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Took 7.68 seconds to spawn the instance on the hypervisor. [ 879.077560] env[61936]: DEBUG nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.078747] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c309a8-a561-4386-8670-f407e69a4ca9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.134293] env[61936]: DEBUG nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 879.158696] env[61936]: DEBUG nova.network.neutron [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Successfully created port: 14553288-1c58-403e-bd50-95de5e72fed1 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.292791] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253031, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.468061] env[61936]: DEBUG nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 879.611967] env[61936]: INFO nova.compute.manager [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Took 27.07 seconds to build instance. [ 879.632315] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18365136-b8fd-4043-bb68-9a3e6401e69b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.642330] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf660d1-dfe6-42e8-a34c-ead011e26502 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.685551] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a941fe4-353c-412e-87b3-d7daef8bea69 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.698028] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06389ece-e6d8-4c30-9eb2-69b3065e6440 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.711031] env[61936]: DEBUG nova.compute.provider_tree [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.778314] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253031, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.741211} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.778893] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.779288] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.779684] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba5ae424-8706-4239-87bd-afe9bc2159a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.803786] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 879.803786] env[61936]: value = "task-1253032" [ 879.803786] env[61936]: _type = "Task" [ 879.803786] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.813330] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253032, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.994278] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.021923] env[61936]: DEBUG nova.network.neutron [-] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.073197] env[61936]: DEBUG nova.compute.manager [req-a6dbbba9-a876-4bf3-9f34-28605097e0a5 req-75bd75ba-9607-4302-8ad2-07073eccf38b service nova] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Received event network-vif-deleted-a5a09509-5534-4354-b48f-c09cfa35b3d6 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 880.113970] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3d1e73af-d427-4566-bc19-5ebcaee1c539 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.152s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.156956] env[61936]: DEBUG nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 880.189969] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 880.190244] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.190392] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 880.190577] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.190723] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 880.190868] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 880.192299] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 880.192579] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 880.192831] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 880.193058] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 880.193299] env[61936]: DEBUG nova.virt.hardware [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 880.194747] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b93141d-d533-449c-bbce-155a55378e1c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.204958] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6945aee3-dbfc-4782-93b2-572b91a642a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.220655] env[61936]: DEBUG nova.scheduler.client.report [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 880.315330] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253032, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069131} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.316805] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.316805] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa6c995-bdf8-49ff-96f8-5ef4b8460083 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.339087] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.340102] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26dd1f34-0dfb-4244-a790-57bbce4e92f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.361078] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 880.361078] env[61936]: value = "task-1253033" [ 880.361078] env[61936]: _type = "Task" [ 880.361078] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.374023] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253033, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.403739] env[61936]: DEBUG nova.network.neutron [-] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.525328] env[61936]: INFO nova.compute.manager [-] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Took 1.71 seconds to deallocate network for instance. [ 880.616019] env[61936]: DEBUG nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 880.620120] env[61936]: DEBUG nova.network.neutron [-] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.725095] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.601s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.726325] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.566s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.726628] env[61936]: DEBUG nova.objects.instance [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lazy-loading 'resources' on Instance uuid e5e3e7af-8104-4f7a-ae18-81e03a932608 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.799536] env[61936]: DEBUG nova.compute.manager [req-23dce868-77db-4e3a-9399-5d5acdf2b6a9 req-95b5ee2a-446c-470d-a3b4-3d129eeb030e service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Received event network-vif-plugged-14553288-1c58-403e-bd50-95de5e72fed1 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 880.800088] env[61936]: DEBUG oslo_concurrency.lockutils [req-23dce868-77db-4e3a-9399-5d5acdf2b6a9 req-95b5ee2a-446c-470d-a3b4-3d129eeb030e service nova] Acquiring lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.800355] env[61936]: DEBUG oslo_concurrency.lockutils [req-23dce868-77db-4e3a-9399-5d5acdf2b6a9 req-95b5ee2a-446c-470d-a3b4-3d129eeb030e service nova] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.800490] env[61936]: DEBUG oslo_concurrency.lockutils [req-23dce868-77db-4e3a-9399-5d5acdf2b6a9 req-95b5ee2a-446c-470d-a3b4-3d129eeb030e service nova] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.800781] env[61936]: DEBUG nova.compute.manager [req-23dce868-77db-4e3a-9399-5d5acdf2b6a9 req-95b5ee2a-446c-470d-a3b4-3d129eeb030e service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] No waiting events found dispatching network-vif-plugged-14553288-1c58-403e-bd50-95de5e72fed1 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 880.801045] env[61936]: WARNING nova.compute.manager [req-23dce868-77db-4e3a-9399-5d5acdf2b6a9 req-95b5ee2a-446c-470d-a3b4-3d129eeb030e service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Received unexpected event network-vif-plugged-14553288-1c58-403e-bd50-95de5e72fed1 for instance with vm_state building and task_state spawning. [ 880.872532] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.894200] env[61936]: DEBUG nova.network.neutron [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Successfully updated port: 14553288-1c58-403e-bd50-95de5e72fed1 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.907879] env[61936]: INFO nova.compute.manager [-] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Took 1.96 seconds to deallocate network for instance. [ 880.955309] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.955604] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.955817] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.956013] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.956223] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.959988] env[61936]: INFO nova.compute.manager [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Terminating instance [ 881.043067] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.043532] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.043780] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.043980] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.044183] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.044390] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.046661] env[61936]: INFO nova.compute.manager [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Terminating instance [ 881.125982] env[61936]: INFO nova.compute.manager [-] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Took 2.40 seconds to deallocate network for instance. [ 881.145999] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.229646] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "ef5376fa-b154-4e78-b577-884050923f89" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.229956] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "ef5376fa-b154-4e78-b577-884050923f89" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.374299] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253033, 'name': ReconfigVM_Task, 'duration_secs': 0.792891} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.377386] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Reconfigured VM instance instance-00000043 to attach disk [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a/aa37292e-8499-46aa-8021-cb8d88b5a35a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.378389] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef3309f0-c4f9-49aa-a85b-2a024615f041 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.388061] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 881.388061] env[61936]: value = "task-1253034" [ 881.388061] env[61936]: _type = "Task" [ 881.388061] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.396863] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "refresh_cache-a8b831c7-bd6a-4218-b19e-1c43e3678c59" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.397129] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquired lock "refresh_cache-a8b831c7-bd6a-4218-b19e-1c43e3678c59" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.397388] env[61936]: DEBUG nova.network.neutron [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.407744] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253034, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.414662] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.465828] env[61936]: DEBUG nova.compute.manager [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 881.466084] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.467230] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc032b62-4926-4c3a-879c-5f9542964905 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.478394] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.481612] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7819302c-d194-4d9d-a51e-256772abfb18 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.489663] env[61936]: DEBUG oslo_vmware.api [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 881.489663] env[61936]: value = "task-1253035" [ 881.489663] env[61936]: _type = "Task" [ 881.489663] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.499662] env[61936]: DEBUG oslo_vmware.api [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.551754] env[61936]: DEBUG nova.compute.manager [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 881.552033] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.553571] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98935ca1-9df4-4774-b4f0-87def3dd0ae7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.563054] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.565913] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be23ad29-0c44-4ac0-b7c1-6abb1c199fe9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.575853] env[61936]: DEBUG oslo_vmware.api [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 881.575853] env[61936]: value = "task-1253036" [ 881.575853] env[61936]: _type = "Task" [ 881.575853] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.588397] env[61936]: DEBUG oslo_vmware.api [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253036, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.608717] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3b9ac0-64f3-425d-b936-5f9d2734bd9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.617369] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643c19fa-3f40-4b24-b626-abaa98ba9e79 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.649065] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.650197] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70b1d91-fee4-4021-af63-825a9b2dac96 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.659327] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565a23d7-ea87-443d-97d1-b903cffdb0e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.674914] env[61936]: DEBUG nova.compute.provider_tree [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.735194] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "ef5376fa-b154-4e78-b577-884050923f89" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.735993] env[61936]: DEBUG nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 881.899584] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253034, 'name': Rename_Task, 'duration_secs': 0.210857} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.901688] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.901966] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a075e5e-9697-46a4-9810-b0b586749664 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.910125] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 881.910125] env[61936]: value = "task-1253037" [ 881.910125] env[61936]: _type = "Task" [ 881.910125] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.919321] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253037, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.931481] env[61936]: DEBUG nova.network.neutron [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.002421] env[61936]: DEBUG oslo_vmware.api [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253035, 'name': PowerOffVM_Task, 'duration_secs': 0.25551} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.003029] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.003226] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.003498] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b18dc761-aa11-4ea4-9902-2a9d5a0ba1e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.090583] env[61936]: DEBUG oslo_vmware.api [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253036, 'name': PowerOffVM_Task, 'duration_secs': 0.24414} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.090583] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.090583] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.090583] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba5a25ee-e214-45e5-b5ef-667b033275ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.094039] env[61936]: DEBUG nova.network.neutron [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Updating instance_info_cache with network_info: [{"id": "14553288-1c58-403e-bd50-95de5e72fed1", "address": "fa:16:3e:cd:e3:8c", "network": {"id": "8ba65d31-038b-4de1-b2bf-f479cb432df2", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-554262829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3202ab76c58f4615bf87a324602770af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14553288-1c", "ovs_interfaceid": "14553288-1c58-403e-bd50-95de5e72fed1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.109875] env[61936]: DEBUG nova.compute.manager [req-abeab672-abdd-47aa-9ced-a90f2cf6c973 req-f783e88d-d40c-443d-98c6-02225061f0fc service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received event network-vif-deleted-cf56feac-af85-4393-84a6-d364545347cc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 882.109986] env[61936]: DEBUG nova.compute.manager [req-abeab672-abdd-47aa-9ced-a90f2cf6c973 req-f783e88d-d40c-443d-98c6-02225061f0fc service nova] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Received event network-vif-deleted-853ddccf-d49b-471f-afc6-5cf28851a036 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 882.110133] env[61936]: DEBUG nova.compute.manager [req-abeab672-abdd-47aa-9ced-a90f2cf6c973 req-f783e88d-d40c-443d-98c6-02225061f0fc service nova] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Received event network-vif-deleted-180a6391-4494-456c-b805-ec2ad1617c73 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 882.120705] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.120968] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.121171] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleting the datastore file [datastore1] efdce04f-b0aa-49a6-94b8-1c960ac33fea {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.121474] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aae5106f-361f-4f24-b176-46a63c50f087 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.130586] env[61936]: DEBUG oslo_vmware.api [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 882.130586] env[61936]: value = "task-1253040" [ 882.130586] env[61936]: _type = "Task" [ 882.130586] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.142644] env[61936]: DEBUG oslo_vmware.api [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.177769] env[61936]: DEBUG nova.scheduler.client.report [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 882.190729] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.190957] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.191162] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleting the datastore file [datastore1] 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.191482] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6662e066-9528-461a-9cec-727419f2db58 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.199910] env[61936]: DEBUG oslo_vmware.api [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 882.199910] env[61936]: value = "task-1253041" [ 882.199910] env[61936]: _type = "Task" [ 882.199910] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.209554] env[61936]: DEBUG oslo_vmware.api [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.243044] env[61936]: DEBUG nova.compute.utils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.244794] env[61936]: DEBUG nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 882.245096] env[61936]: DEBUG nova.network.neutron [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 882.290305] env[61936]: DEBUG nova.policy [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d93484519954f87b9d95d04eccb0389', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3edfcb5373584d0ca431e3783aea86cd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 882.421610] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253037, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.566891] env[61936]: DEBUG nova.network.neutron [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Successfully created port: 464d9afb-54f5-47c5-aedf-ac53cb44d6b9 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.600019] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Releasing lock "refresh_cache-a8b831c7-bd6a-4218-b19e-1c43e3678c59" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.600019] env[61936]: DEBUG nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Instance network_info: |[{"id": "14553288-1c58-403e-bd50-95de5e72fed1", "address": "fa:16:3e:cd:e3:8c", "network": {"id": "8ba65d31-038b-4de1-b2bf-f479cb432df2", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-554262829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3202ab76c58f4615bf87a324602770af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14553288-1c", "ovs_interfaceid": "14553288-1c58-403e-bd50-95de5e72fed1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 882.600019] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:e3:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7bf7d4-8e0c-4cee-84ba-244e73ef6379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14553288-1c58-403e-bd50-95de5e72fed1', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.606190] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Creating folder: Project (3202ab76c58f4615bf87a324602770af). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.606671] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7c3c4607-02e0-4432-ba16-dd4396de6ee3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.623082] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Created folder: Project (3202ab76c58f4615bf87a324602770af) in parent group-v269874. [ 882.623082] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Creating folder: Instances. Parent ref: group-v269963. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.623082] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2cd89075-9f55-441b-94d5-7541241a4614 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.637808] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Created folder: Instances in parent group-v269963. [ 882.638070] env[61936]: DEBUG oslo.service.loopingcall [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.638680] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.638898] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-191dfa89-0c5c-4634-bd5a-2d316e7cb7bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.656236] env[61936]: DEBUG oslo_vmware.api [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187263} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.656974] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.657929] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.657929] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.657929] env[61936]: INFO nova.compute.manager [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Took 1.19 seconds to destroy the instance on the hypervisor. [ 882.657929] env[61936]: DEBUG oslo.service.loopingcall [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.657929] env[61936]: DEBUG nova.compute.manager [-] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 882.658176] env[61936]: DEBUG nova.network.neutron [-] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.663617] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.663617] env[61936]: value = "task-1253044" [ 882.663617] env[61936]: _type = "Task" [ 882.663617] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.672086] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253044, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.685977] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.960s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.692692] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.974s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.694825] env[61936]: INFO nova.compute.claims [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.715569] env[61936]: DEBUG oslo_vmware.api [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253041, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169746} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.716662] env[61936]: INFO nova.scheduler.client.report [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Deleted allocations for instance e5e3e7af-8104-4f7a-ae18-81e03a932608 [ 882.717815] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.718117] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.720765] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.720765] env[61936]: INFO nova.compute.manager [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 882.720765] env[61936]: DEBUG oslo.service.loopingcall [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.721528] env[61936]: DEBUG nova.compute.manager [-] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 882.721528] env[61936]: DEBUG nova.network.neutron [-] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.749175] env[61936]: DEBUG nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 882.834462] env[61936]: DEBUG nova.compute.manager [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Received event network-changed-14553288-1c58-403e-bd50-95de5e72fed1 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 882.834462] env[61936]: DEBUG nova.compute.manager [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Refreshing instance network info cache due to event network-changed-14553288-1c58-403e-bd50-95de5e72fed1. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 882.834462] env[61936]: DEBUG oslo_concurrency.lockutils [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] Acquiring lock "refresh_cache-a8b831c7-bd6a-4218-b19e-1c43e3678c59" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.834462] env[61936]: DEBUG oslo_concurrency.lockutils [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] Acquired lock "refresh_cache-a8b831c7-bd6a-4218-b19e-1c43e3678c59" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.834462] env[61936]: DEBUG nova.network.neutron [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Refreshing network info cache for port 14553288-1c58-403e-bd50-95de5e72fed1 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.927541] env[61936]: DEBUG oslo_vmware.api [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253037, 'name': PowerOnVM_Task, 'duration_secs': 0.596404} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.927541] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.927541] env[61936]: DEBUG nova.compute.manager [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 882.927541] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d2c9b1-7ccc-4445-85d4-d17d6bbc1fed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.179249] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253044, 'name': CreateVM_Task, 'duration_secs': 0.360701} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.180137] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 883.181074] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.181412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.181910] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.182391] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3fce5d7-a116-46a4-a357-4be91f571728 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.189623] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 883.189623] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ef9ca2-5d39-3807-379f-17a9e87dfea9" [ 883.189623] env[61936]: _type = "Task" [ 883.189623] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.204092] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ef9ca2-5d39-3807-379f-17a9e87dfea9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.230183] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cb05aaa4-a952-48a6-a847-3c9c7ef30c36 tempest-ServerTagsTestJSON-134261133 tempest-ServerTagsTestJSON-134261133-project-member] Lock "e5e3e7af-8104-4f7a-ae18-81e03a932608" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.570s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.461713] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.701526] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ef9ca2-5d39-3807-379f-17a9e87dfea9, 'name': SearchDatastore_Task, 'duration_secs': 0.010742} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.701975] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.702124] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.702371] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.702519] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.702699] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.702975] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a607293-2fb1-4651-87d8-a7c7f3f4ab7c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.719138] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.719138] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.719980] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0f7cb8f-c3e3-436e-a35b-b5fd33619884 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.727100] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 883.727100] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520ed0c3-d130-82aa-ebb9-6d4ae29c7889" [ 883.727100] env[61936]: _type = "Task" [ 883.727100] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.735567] env[61936]: DEBUG nova.network.neutron [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Updated VIF entry in instance network info cache for port 14553288-1c58-403e-bd50-95de5e72fed1. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 883.735951] env[61936]: DEBUG nova.network.neutron [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Updating instance_info_cache with network_info: [{"id": "14553288-1c58-403e-bd50-95de5e72fed1", "address": "fa:16:3e:cd:e3:8c", "network": {"id": "8ba65d31-038b-4de1-b2bf-f479cb432df2", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-554262829-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3202ab76c58f4615bf87a324602770af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14553288-1c", "ovs_interfaceid": "14553288-1c58-403e-bd50-95de5e72fed1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.746018] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520ed0c3-d130-82aa-ebb9-6d4ae29c7889, 'name': SearchDatastore_Task, 'duration_secs': 0.012062} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.746309] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38f31095-22d6-44ec-a5ce-f854f6521afd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.753346] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 883.753346] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ae8537-4c61-6a62-d6c7-965c6da815fc" [ 883.753346] env[61936]: _type = "Task" [ 883.753346] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.763394] env[61936]: DEBUG nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 883.771836] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ae8537-4c61-6a62-d6c7-965c6da815fc, 'name': SearchDatastore_Task, 'duration_secs': 0.011148} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.775268] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.775517] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] a8b831c7-bd6a-4218-b19e-1c43e3678c59/a8b831c7-bd6a-4218-b19e-1c43e3678c59.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.777170] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7ea9d3a-63a0-439c-8316-0dfed89a28f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.790769] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 883.790769] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.790769] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 883.790769] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.791211] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 883.791211] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 883.791483] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 883.791612] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 883.791780] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 883.791938] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 883.792119] env[61936]: DEBUG nova.virt.hardware [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 883.793700] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56942311-a47b-4bf4-b821-a6466e47c60a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.801617] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 883.801617] env[61936]: value = "task-1253045" [ 883.801617] env[61936]: _type = "Task" [ 883.801617] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.810569] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdf9f37-cc2e-4372-8c18-27fa6a24e3c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.821074] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.853021] env[61936]: DEBUG nova.network.neutron [-] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.964784] env[61936]: DEBUG nova.network.neutron [-] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.135028] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6dd6d5-f4b8-4bd3-9ed9-178e8771092c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.144757] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9202e691-3647-4b4e-bdd3-95bde969dea1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.152553] env[61936]: DEBUG nova.compute.manager [req-9553e2bf-b1b2-489c-8ea9-a2bc35191df9 req-1ac6b730-e659-4220-aaef-74e5f447e960 service nova] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Received event network-vif-deleted-dc5d4628-264f-4f4e-9006-834caeb7b139 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 884.152553] env[61936]: DEBUG nova.compute.manager [req-9553e2bf-b1b2-489c-8ea9-a2bc35191df9 req-1ac6b730-e659-4220-aaef-74e5f447e960 service nova] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Received event network-vif-deleted-2a56df4a-9c49-4ffb-8c78-bdd4edd9714a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 884.197755] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f111a2-5baa-41ea-9985-3bfc6b75c657 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.208475] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e904d8-6340-48fd-b354-2e6096d7b0eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.226088] env[61936]: DEBUG nova.compute.provider_tree [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.238583] env[61936]: DEBUG oslo_concurrency.lockutils [req-e6fa865c-65dc-49a0-bd5c-282022f37b48 req-a0b9425b-fb25-4213-91e1-71202252f86b service nova] Releasing lock "refresh_cache-a8b831c7-bd6a-4218-b19e-1c43e3678c59" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.285831] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "aa37292e-8499-46aa-8021-cb8d88b5a35a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.286197] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "aa37292e-8499-46aa-8021-cb8d88b5a35a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.286404] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "aa37292e-8499-46aa-8021-cb8d88b5a35a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.286596] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "aa37292e-8499-46aa-8021-cb8d88b5a35a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.286770] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "aa37292e-8499-46aa-8021-cb8d88b5a35a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.289273] env[61936]: INFO nova.compute.manager [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Terminating instance [ 884.317315] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510199} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.318473] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] a8b831c7-bd6a-4218-b19e-1c43e3678c59/a8b831c7-bd6a-4218-b19e-1c43e3678c59.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 884.318773] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.319423] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b15ec5b-d045-43b6-8c34-f2b484eb8353 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.329720] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 884.329720] env[61936]: value = "task-1253046" [ 884.329720] env[61936]: _type = "Task" [ 884.329720] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.338908] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.355648] env[61936]: INFO nova.compute.manager [-] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Took 1.70 seconds to deallocate network for instance. [ 884.467207] env[61936]: INFO nova.compute.manager [-] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Took 1.75 seconds to deallocate network for instance. [ 884.644240] env[61936]: DEBUG nova.network.neutron [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Successfully updated port: 464d9afb-54f5-47c5-aedf-ac53cb44d6b9 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.729913] env[61936]: DEBUG nova.scheduler.client.report [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 884.793265] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "refresh_cache-aa37292e-8499-46aa-8021-cb8d88b5a35a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.793481] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "refresh_cache-aa37292e-8499-46aa-8021-cb8d88b5a35a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.793668] env[61936]: DEBUG nova.network.neutron [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.845528] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092975} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.845956] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.847711] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c837277f-b1b0-4032-8f71-65cf0a5e8480 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.869811] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.882951] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] a8b831c7-bd6a-4218-b19e-1c43e3678c59/a8b831c7-bd6a-4218-b19e-1c43e3678c59.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.883709] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a257ed0-2c7b-4dbc-9a07-2e3bb218f4b6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.903092] env[61936]: DEBUG nova.compute.manager [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Received event network-vif-plugged-464d9afb-54f5-47c5-aedf-ac53cb44d6b9 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 884.903092] env[61936]: DEBUG oslo_concurrency.lockutils [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] Acquiring lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.903092] env[61936]: DEBUG oslo_concurrency.lockutils [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.903092] env[61936]: DEBUG oslo_concurrency.lockutils [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.903092] env[61936]: DEBUG nova.compute.manager [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] No waiting events found dispatching network-vif-plugged-464d9afb-54f5-47c5-aedf-ac53cb44d6b9 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 884.903092] env[61936]: WARNING nova.compute.manager [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Received unexpected event network-vif-plugged-464d9afb-54f5-47c5-aedf-ac53cb44d6b9 for instance with vm_state building and task_state spawning. [ 884.903092] env[61936]: DEBUG nova.compute.manager [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Received event network-changed-464d9afb-54f5-47c5-aedf-ac53cb44d6b9 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 884.903092] env[61936]: DEBUG nova.compute.manager [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Refreshing instance network info cache due to event network-changed-464d9afb-54f5-47c5-aedf-ac53cb44d6b9. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 884.903092] env[61936]: DEBUG oslo_concurrency.lockutils [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] Acquiring lock "refresh_cache-3af90345-f5bf-4d25-90ac-c5389aef5c8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.903092] env[61936]: DEBUG oslo_concurrency.lockutils [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] Acquired lock "refresh_cache-3af90345-f5bf-4d25-90ac-c5389aef5c8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.903494] env[61936]: DEBUG nova.network.neutron [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Refreshing network info cache for port 464d9afb-54f5-47c5-aedf-ac53cb44d6b9 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.911946] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 884.911946] env[61936]: value = "task-1253047" [ 884.911946] env[61936]: _type = "Task" [ 884.911946] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.929031] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253047, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.976672] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.149716] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "refresh_cache-3af90345-f5bf-4d25-90ac-c5389aef5c8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.236243] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.236975] env[61936]: DEBUG nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 885.239940] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.246s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.241606] env[61936]: INFO nova.compute.claims [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.315419] env[61936]: DEBUG nova.network.neutron [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.384302] env[61936]: DEBUG nova.network.neutron [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.424467] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253047, 'name': ReconfigVM_Task, 'duration_secs': 0.361112} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.424833] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Reconfigured VM instance instance-00000046 to attach disk [datastore1] a8b831c7-bd6a-4218-b19e-1c43e3678c59/a8b831c7-bd6a-4218-b19e-1c43e3678c59.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.425569] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-417a2e00-4ecf-4f88-b495-4b933f1d2f61 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.434561] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 885.434561] env[61936]: value = "task-1253048" [ 885.434561] env[61936]: _type = "Task" [ 885.434561] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.438968] env[61936]: DEBUG nova.network.neutron [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 885.448055] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253048, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.534543] env[61936]: DEBUG nova.network.neutron [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.752020] env[61936]: DEBUG nova.compute.utils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.752020] env[61936]: DEBUG nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 885.752020] env[61936]: DEBUG nova.network.neutron [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.794486] env[61936]: DEBUG nova.policy [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a346002c214e4b238f323ffffc6b61b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c334e9711bc64304ab017eb6ab3e2d23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.887181] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "refresh_cache-aa37292e-8499-46aa-8021-cb8d88b5a35a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.887634] env[61936]: DEBUG nova.compute.manager [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 885.887824] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.888804] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f916693d-2daf-4657-ae22-7d45bfd38d50 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.898026] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.898304] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e73f1d10-0901-4fc3-9c5d-d1e81ace38a0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.906971] env[61936]: DEBUG oslo_vmware.api [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 885.906971] env[61936]: value = "task-1253049" [ 885.906971] env[61936]: _type = "Task" [ 885.906971] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.915817] env[61936]: DEBUG oslo_vmware.api [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253049, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.947702] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253048, 'name': Rename_Task, 'duration_secs': 0.185674} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.948150] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.948454] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c34082a7-a951-48ba-a9c2-9ba8e89e006a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.957160] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 885.957160] env[61936]: value = "task-1253050" [ 885.957160] env[61936]: _type = "Task" [ 885.957160] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.969049] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253050, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.038502] env[61936]: DEBUG oslo_concurrency.lockutils [req-a27d64c9-5ea9-4fef-aa74-486531d4887d req-732eee6b-29e3-4844-8866-8b4a98f5dd5d service nova] Releasing lock "refresh_cache-3af90345-f5bf-4d25-90ac-c5389aef5c8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.038917] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquired lock "refresh_cache-3af90345-f5bf-4d25-90ac-c5389aef5c8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.039098] env[61936]: DEBUG nova.network.neutron [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.074957] env[61936]: DEBUG nova.network.neutron [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Successfully created port: 9f132878-bd3a-45e4-a5a0-03051b7bfd0e {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.257153] env[61936]: DEBUG nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 886.418604] env[61936]: DEBUG oslo_vmware.api [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253049, 'name': PowerOffVM_Task, 'duration_secs': 0.172475} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.418992] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.419242] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.419555] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-658f181b-ffd6-4f8b-8f0a-2f4a7d942819 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.457289] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.457631] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.457755] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleting the datastore file [datastore2] aa37292e-8499-46aa-8021-cb8d88b5a35a {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.458074] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c38a0f97-070a-45f5-84b3-f028c2625b53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.470214] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253050, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.474361] env[61936]: DEBUG oslo_vmware.api [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 886.474361] env[61936]: value = "task-1253052" [ 886.474361] env[61936]: _type = "Task" [ 886.474361] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.484920] env[61936]: DEBUG oslo_vmware.api [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253052, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.584692] env[61936]: DEBUG nova.network.neutron [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.616984] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb91c74-7d7c-4e98-9d9a-dc6c02879525 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.627376] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d76531-33c9-47f3-a06b-98fbb376e582 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.659480] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9566a3bd-0daf-4c12-af08-a2fc361aaad7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.667910] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ebe9f0-4844-43ba-9223-107be0eab23d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.682122] env[61936]: DEBUG nova.compute.provider_tree [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.876783] env[61936]: DEBUG nova.network.neutron [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Updating instance_info_cache with network_info: [{"id": "464d9afb-54f5-47c5-aedf-ac53cb44d6b9", "address": "fa:16:3e:ce:f3:d9", "network": {"id": "6fff04ed-ad88-4a0f-8f27-1f173854fa54", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-765060802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3edfcb5373584d0ca431e3783aea86cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap464d9afb-54", "ovs_interfaceid": "464d9afb-54f5-47c5-aedf-ac53cb44d6b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.972050] env[61936]: DEBUG oslo_vmware.api [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253050, 'name': PowerOnVM_Task, 'duration_secs': 0.53455} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.972382] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.972643] env[61936]: INFO nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Took 6.82 seconds to spawn the instance on the hypervisor. [ 886.972861] env[61936]: DEBUG nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.973782] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8286b121-2d3c-44a3-8be4-2dc39d292634 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.989845] env[61936]: DEBUG oslo_vmware.api [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253052, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122764} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.990355] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.990628] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.990853] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.991065] env[61936]: INFO nova.compute.manager [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 886.991383] env[61936]: DEBUG oslo.service.loopingcall [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.991929] env[61936]: DEBUG nova.compute.manager [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 886.991929] env[61936]: DEBUG nova.network.neutron [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 887.015672] env[61936]: DEBUG nova.network.neutron [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.138289] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d4a9a-bfa3-e077-c5ee-528705949da0/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 887.139342] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24402a4-f608-4a73-b3ca-1b9b398eaf60 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.146226] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d4a9a-bfa3-e077-c5ee-528705949da0/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 887.146414] env[61936]: ERROR oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d4a9a-bfa3-e077-c5ee-528705949da0/disk-0.vmdk due to incomplete transfer. [ 887.146642] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cb38a0ba-e427-4964-8a6d-5e68208d9383 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.154262] env[61936]: DEBUG oslo_vmware.rw_handles [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529d4a9a-bfa3-e077-c5ee-528705949da0/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 887.154490] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Uploaded image 258fa9b9-a92c-4360-8f49-b8b720244f4e to the Glance image server {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 887.156825] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Destroying the VM {{(pid=61936) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 887.157409] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ba2d2b8a-c6bf-405e-8d2d-9468145f1c2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.163805] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 887.163805] env[61936]: value = "task-1253053" [ 887.163805] env[61936]: _type = "Task" [ 887.163805] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.171943] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253053, 'name': Destroy_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.186986] env[61936]: DEBUG nova.scheduler.client.report [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 887.272457] env[61936]: DEBUG nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 887.293406] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 887.293622] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.293785] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 887.293973] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.294195] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 887.294294] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 887.294504] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 887.294663] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 887.294829] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 887.295033] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 887.295180] env[61936]: DEBUG nova.virt.hardware [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 887.296094] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740cc3a6-63c4-4394-b0ac-5b39ece034e5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.304690] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f409d4f-0c29-4bac-a6d1-e25af0116fcd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.380095] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Releasing lock "refresh_cache-3af90345-f5bf-4d25-90ac-c5389aef5c8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.380236] env[61936]: DEBUG nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Instance network_info: |[{"id": "464d9afb-54f5-47c5-aedf-ac53cb44d6b9", "address": "fa:16:3e:ce:f3:d9", "network": {"id": "6fff04ed-ad88-4a0f-8f27-1f173854fa54", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-765060802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3edfcb5373584d0ca431e3783aea86cd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "34a581cb-6d33-4e2e-af50-735a6749d6da", "external-id": "nsx-vlan-transportzone-673", "segmentation_id": 673, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap464d9afb-54", "ovs_interfaceid": "464d9afb-54f5-47c5-aedf-ac53cb44d6b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 887.380698] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:f3:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '34a581cb-6d33-4e2e-af50-735a6749d6da', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '464d9afb-54f5-47c5-aedf-ac53cb44d6b9', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.388322] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Creating folder: Project (3edfcb5373584d0ca431e3783aea86cd). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 887.388610] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a03e479-122a-4275-b364-30307fbfdd76 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.401201] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Created folder: Project (3edfcb5373584d0ca431e3783aea86cd) in parent group-v269874. [ 887.401490] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Creating folder: Instances. Parent ref: group-v269966. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 887.401794] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99b86910-57b8-4a72-b014-a6b981ee16e6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.412675] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Created folder: Instances in parent group-v269966. [ 887.412675] env[61936]: DEBUG oslo.service.loopingcall [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.412675] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.412675] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e7c5f1a-6444-4bc5-80d2-df95ac9d004e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.432481] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.432481] env[61936]: value = "task-1253056" [ 887.432481] env[61936]: _type = "Task" [ 887.432481] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.440875] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253056, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.500334] env[61936]: INFO nova.compute.manager [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Took 29.61 seconds to build instance. [ 887.507321] env[61936]: DEBUG nova.compute.manager [req-aed8c45a-71e9-4b60-845b-145f4695be97 req-86af40dd-4f8a-4dd3-990f-c30c2e6bf3ad service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Received event network-vif-plugged-9f132878-bd3a-45e4-a5a0-03051b7bfd0e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 887.508674] env[61936]: DEBUG oslo_concurrency.lockutils [req-aed8c45a-71e9-4b60-845b-145f4695be97 req-86af40dd-4f8a-4dd3-990f-c30c2e6bf3ad service nova] Acquiring lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.508674] env[61936]: DEBUG oslo_concurrency.lockutils [req-aed8c45a-71e9-4b60-845b-145f4695be97 req-86af40dd-4f8a-4dd3-990f-c30c2e6bf3ad service nova] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.508674] env[61936]: DEBUG oslo_concurrency.lockutils [req-aed8c45a-71e9-4b60-845b-145f4695be97 req-86af40dd-4f8a-4dd3-990f-c30c2e6bf3ad service nova] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.508674] env[61936]: DEBUG nova.compute.manager [req-aed8c45a-71e9-4b60-845b-145f4695be97 req-86af40dd-4f8a-4dd3-990f-c30c2e6bf3ad service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] No waiting events found dispatching network-vif-plugged-9f132878-bd3a-45e4-a5a0-03051b7bfd0e {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 887.508674] env[61936]: WARNING nova.compute.manager [req-aed8c45a-71e9-4b60-845b-145f4695be97 req-86af40dd-4f8a-4dd3-990f-c30c2e6bf3ad service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Received unexpected event network-vif-plugged-9f132878-bd3a-45e4-a5a0-03051b7bfd0e for instance with vm_state building and task_state spawning. [ 887.518452] env[61936]: DEBUG nova.network.neutron [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.608888] env[61936]: DEBUG nova.network.neutron [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Successfully updated port: 9f132878-bd3a-45e4-a5a0-03051b7bfd0e {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.675304] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253053, 'name': Destroy_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.691703] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.692252] env[61936]: DEBUG nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 887.695055] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.799s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.696360] env[61936]: INFO nova.compute.claims [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.943552] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253056, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.010493] env[61936]: DEBUG oslo_concurrency.lockutils [None req-310c945b-1ca2-48b8-b5a6-e69e81ace3fd tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.192s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.020699] env[61936]: INFO nova.compute.manager [-] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Took 1.03 seconds to deallocate network for instance. [ 888.111858] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "refresh_cache-f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.111991] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "refresh_cache-f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.112162] env[61936]: DEBUG nova.network.neutron [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.176537] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253053, 'name': Destroy_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.201034] env[61936]: DEBUG nova.compute.utils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.206032] env[61936]: DEBUG nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 888.206032] env[61936]: DEBUG nova.network.neutron [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 888.261045] env[61936]: DEBUG nova.policy [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcee2be71ae044338df4422a3c575ca4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e81b119251964bdb8e5a0ea84b29f2a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.443591] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253056, 'name': CreateVM_Task, 'duration_secs': 0.858554} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.443726] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.444784] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.444962] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.445306] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.445575] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ab95172-7a44-4276-a3ef-f8733b9cc169 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.451938] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 888.451938] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5229fd1f-1697-3d9b-aeaa-7b54598253be" [ 888.451938] env[61936]: _type = "Task" [ 888.451938] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.462389] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5229fd1f-1697-3d9b-aeaa-7b54598253be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.513833] env[61936]: DEBUG nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 888.531053] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.606997] env[61936]: DEBUG nova.network.neutron [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Successfully created port: 39c4ceec-fef3-4d7d-8a6b-27fc061802a0 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.664796] env[61936]: DEBUG nova.network.neutron [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.676705] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253053, 'name': Destroy_Task, 'duration_secs': 1.371277} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.676971] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Destroyed the VM [ 888.677333] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Deleting Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 888.677599] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-316ad64e-af1d-4418-8f0b-3f612228a261 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.685473] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 888.685473] env[61936]: value = "task-1253057" [ 888.685473] env[61936]: _type = "Task" [ 888.685473] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.694044] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253057, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.709384] env[61936]: DEBUG nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 888.862217] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.862739] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.863355] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.863791] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.864108] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.866440] env[61936]: INFO nova.compute.manager [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Terminating instance [ 888.919489] env[61936]: DEBUG nova.network.neutron [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Updating instance_info_cache with network_info: [{"id": "9f132878-bd3a-45e4-a5a0-03051b7bfd0e", "address": "fa:16:3e:7d:8e:75", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f132878-bd", "ovs_interfaceid": "9f132878-bd3a-45e4-a5a0-03051b7bfd0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.963714] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5229fd1f-1697-3d9b-aeaa-7b54598253be, 'name': SearchDatastore_Task, 'duration_secs': 0.013485} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.966724] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.967036] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.967319] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.967511] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.967740] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.968312] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8353d76-854e-4339-bacd-c101692b02f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.979492] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.979696] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.983344] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31cfd5e2-62b3-4f75-8ecf-9bfa19e64547 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.992066] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 888.992066] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52792727-7a17-24ca-9bf0-629cd0150905" [ 888.992066] env[61936]: _type = "Task" [ 888.992066] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.009134] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52792727-7a17-24ca-9bf0-629cd0150905, 'name': SearchDatastore_Task, 'duration_secs': 0.011826} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.009134] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0b84640-5506-42e6-97c0-6cc3ffe21a54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.015341] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 889.015341] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5261dd26-ee48-f44c-2ffc-7270c28452f4" [ 889.015341] env[61936]: _type = "Task" [ 889.015341] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.034765] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5261dd26-ee48-f44c-2ffc-7270c28452f4, 'name': SearchDatastore_Task, 'duration_secs': 0.010716} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.034765] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.034943] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 3af90345-f5bf-4d25-90ac-c5389aef5c8e/3af90345-f5bf-4d25-90ac-c5389aef5c8e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.035437] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2cb98a32-2db0-452c-9ab1-e4f719777f4c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.043177] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.048021] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 889.048021] env[61936]: value = "task-1253058" [ 889.048021] env[61936]: _type = "Task" [ 889.048021] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.055579] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253058, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.110434] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240c0c76-1285-468b-b1b4-0362da8850cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.118998] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68fc0c06-3022-4320-b857-c18d91ed5db7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.152034] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ec373a-1eef-45dd-ade8-eb93df26297a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.161036] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecdd222-2f85-497a-a34e-1b6df65e9fe3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.177916] env[61936]: DEBUG nova.compute.provider_tree [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.196396] env[61936]: DEBUG oslo_vmware.api [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253057, 'name': RemoveSnapshot_Task, 'duration_secs': 0.483724} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.196679] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Deleted Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 889.196948] env[61936]: INFO nova.compute.manager [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Took 15.41 seconds to snapshot the instance on the hypervisor. [ 889.373792] env[61936]: DEBUG nova.compute.manager [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 889.374322] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.375858] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c8d2d3-40ca-4e12-8e05-1838c382a77c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.388146] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.388681] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0fd9b00f-c754-4258-968c-0fa38ac64cd3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.422964] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "refresh_cache-f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.423292] env[61936]: DEBUG nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Instance network_info: |[{"id": "9f132878-bd3a-45e4-a5a0-03051b7bfd0e", "address": "fa:16:3e:7d:8e:75", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f132878-bd", "ovs_interfaceid": "9f132878-bd3a-45e4-a5a0-03051b7bfd0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 889.423743] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:8e:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a44713-0af1-486e-bc0d-00e03a769fa4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f132878-bd3a-45e4-a5a0-03051b7bfd0e', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.432018] env[61936]: DEBUG oslo.service.loopingcall [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.432455] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.432715] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd3d144d-72e1-4092-b049-dfb39289370e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.456167] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.456167] env[61936]: value = "task-1253060" [ 889.456167] env[61936]: _type = "Task" [ 889.456167] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.466246] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253060, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.495517] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.495925] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.496059] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleting the datastore file [datastore2] 18175943-35d2-4792-a8ea-e54b4ffbaf2f {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.496400] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-459fa7d2-8a73-4e51-930f-f644e02dd081 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.505862] env[61936]: DEBUG oslo_vmware.api [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 889.505862] env[61936]: value = "task-1253061" [ 889.505862] env[61936]: _type = "Task" [ 889.505862] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.516352] env[61936]: DEBUG oslo_vmware.api [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253061, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.532623] env[61936]: DEBUG nova.compute.manager [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Received event network-changed-9f132878-bd3a-45e4-a5a0-03051b7bfd0e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 889.532851] env[61936]: DEBUG nova.compute.manager [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Refreshing instance network info cache due to event network-changed-9f132878-bd3a-45e4-a5a0-03051b7bfd0e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 889.533077] env[61936]: DEBUG oslo_concurrency.lockutils [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] Acquiring lock "refresh_cache-f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.533228] env[61936]: DEBUG oslo_concurrency.lockutils [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] Acquired lock "refresh_cache-f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.533401] env[61936]: DEBUG nova.network.neutron [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Refreshing network info cache for port 9f132878-bd3a-45e4-a5a0-03051b7bfd0e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.558996] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253058, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488574} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.558996] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 3af90345-f5bf-4d25-90ac-c5389aef5c8e/3af90345-f5bf-4d25-90ac-c5389aef5c8e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.559230] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.559536] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1d4d748-d522-48d3-8450-e485f96ba762 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.567639] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 889.567639] env[61936]: value = "task-1253062" [ 889.567639] env[61936]: _type = "Task" [ 889.567639] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.578275] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.624939] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "b780470b-57d0-43e5-be52-539c78cada6b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.625224] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "b780470b-57d0-43e5-be52-539c78cada6b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.681161] env[61936]: DEBUG nova.scheduler.client.report [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.703829] env[61936]: DEBUG nova.compute.manager [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Instance disappeared during snapshot {{(pid=61936) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 889.719441] env[61936]: DEBUG nova.compute.manager [None req-0fbc40f4-4ccb-4ac4-85d9-cd98bb301844 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image not found during clean up 258fa9b9-a92c-4360-8f49-b8b720244f4e {{(pid=61936) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4600}} [ 889.720956] env[61936]: DEBUG nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 889.747534] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 889.747958] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.748271] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 889.748624] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.748883] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 889.749152] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 889.749560] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 889.749866] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 889.750221] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 889.750565] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 889.750896] env[61936]: DEBUG nova.virt.hardware [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 889.752837] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e333e80d-78a4-4901-833d-82fcffdd521c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.765591] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d2e55a-9436-49bd-8a67-b955614a4bae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.967766] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253060, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.016514] env[61936]: DEBUG oslo_vmware.api [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253061, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.462713} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.017096] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.017096] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.017344] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.017505] env[61936]: INFO nova.compute.manager [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Took 0.64 seconds to destroy the instance on the hypervisor. [ 890.017942] env[61936]: DEBUG oslo.service.loopingcall [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.018019] env[61936]: DEBUG nova.compute.manager [-] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 890.018104] env[61936]: DEBUG nova.network.neutron [-] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.078751] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079933} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.079134] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.080129] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03744a99-8efc-4da2-9a95-2ce31778253b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.097264] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.097728] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.098040] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.098282] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.098474] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.109219] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 3af90345-f5bf-4d25-90ac-c5389aef5c8e/3af90345-f5bf-4d25-90ac-c5389aef5c8e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.112129] env[61936]: INFO nova.compute.manager [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Terminating instance [ 890.113689] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15f38326-2557-4483-8c86-713a12e60780 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.128405] env[61936]: DEBUG nova.network.neutron [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Successfully updated port: 39c4ceec-fef3-4d7d-8a6b-27fc061802a0 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.138430] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 890.138430] env[61936]: value = "task-1253063" [ 890.138430] env[61936]: _type = "Task" [ 890.138430] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.151276] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253063, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.186167] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.186728] env[61936]: DEBUG nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 890.189732] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.731s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.191230] env[61936]: INFO nova.compute.claims [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.468909] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253060, 'name': CreateVM_Task, 'duration_secs': 0.663222} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.469253] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.469960] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.470148] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.470471] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.470773] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3d4fb28-ca91-4632-9659-8b9a69929f36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.475439] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 890.475439] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52980217-ca42-8f9a-1340-643fed1372cc" [ 890.475439] env[61936]: _type = "Task" [ 890.475439] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.485375] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52980217-ca42-8f9a-1340-643fed1372cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.632550] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-628882ff-2f0f-49ae-8454-d3791fb55545" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.632748] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-628882ff-2f0f-49ae-8454-d3791fb55545" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.632978] env[61936]: DEBUG nova.network.neutron [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.634611] env[61936]: DEBUG nova.compute.manager [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 890.634849] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 890.635712] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed683e1e-8004-4033-9c9b-7d6df4b2c37c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.650972] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253063, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.654971] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.656020] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-082fe367-5a9f-4de9-9c94-b9195a0bb888 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.667354] env[61936]: DEBUG oslo_vmware.api [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 890.667354] env[61936]: value = "task-1253064" [ 890.667354] env[61936]: _type = "Task" [ 890.667354] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.678056] env[61936]: DEBUG oslo_vmware.api [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253064, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.696264] env[61936]: DEBUG nova.compute.utils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.703022] env[61936]: DEBUG nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 890.703022] env[61936]: DEBUG nova.network.neutron [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 890.740312] env[61936]: DEBUG nova.network.neutron [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Updated VIF entry in instance network info cache for port 9f132878-bd3a-45e4-a5a0-03051b7bfd0e. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.740761] env[61936]: DEBUG nova.network.neutron [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Updating instance_info_cache with network_info: [{"id": "9f132878-bd3a-45e4-a5a0-03051b7bfd0e", "address": "fa:16:3e:7d:8e:75", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f132878-bd", "ovs_interfaceid": "9f132878-bd3a-45e4-a5a0-03051b7bfd0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.743340] env[61936]: DEBUG nova.policy [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b05bdcace008487dba580405eb672ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90c29449db6c4beaa7f190d9225cb08a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 890.887474] env[61936]: DEBUG nova.compute.manager [req-70df2aa9-c4de-4a91-9924-8f5228d8d1f3 req-747cbb47-9490-46f1-b65e-2ea131ccb305 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Received event network-vif-deleted-b7963334-7c72-41c3-9e98-72a837dbb33f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 890.887776] env[61936]: INFO nova.compute.manager [req-70df2aa9-c4de-4a91-9924-8f5228d8d1f3 req-747cbb47-9490-46f1-b65e-2ea131ccb305 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Neutron deleted interface b7963334-7c72-41c3-9e98-72a837dbb33f; detaching it from the instance and deleting it from the info cache [ 890.888531] env[61936]: DEBUG nova.network.neutron [req-70df2aa9-c4de-4a91-9924-8f5228d8d1f3 req-747cbb47-9490-46f1-b65e-2ea131ccb305 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.987039] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52980217-ca42-8f9a-1340-643fed1372cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010266} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.987545] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.987623] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.987855] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.988011] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.988237] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.988496] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58b95a24-8ff7-4ec8-9130-b76bc0c8e07d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.998476] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.998672] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.999497] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dd44cce-6d94-49e2-b621-2fe41dcf974a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.005239] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 891.005239] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520a33c7-8005-55f8-581d-66de3632605a" [ 891.005239] env[61936]: _type = "Task" [ 891.005239] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.014016] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520a33c7-8005-55f8-581d-66de3632605a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.015499] env[61936]: DEBUG nova.network.neutron [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Successfully created port: 691101d8-d631-42c1-80ce-2aba05214046 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 891.152656] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253063, 'name': ReconfigVM_Task, 'duration_secs': 0.843621} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.152952] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 3af90345-f5bf-4d25-90ac-c5389aef5c8e/3af90345-f5bf-4d25-90ac-c5389aef5c8e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.153609] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa1c6715-3d09-40e9-b30b-260fde1bbe3a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.161216] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 891.161216] env[61936]: value = "task-1253065" [ 891.161216] env[61936]: _type = "Task" [ 891.161216] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.169919] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253065, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.170820] env[61936]: DEBUG nova.network.neutron [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.180944] env[61936]: DEBUG oslo_vmware.api [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253064, 'name': PowerOffVM_Task, 'duration_secs': 0.252729} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.181271] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.181406] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.181661] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1567cba8-e6d2-4b31-9a3b-1a63ced29b83 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.200134] env[61936]: DEBUG nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 891.246164] env[61936]: DEBUG oslo_concurrency.lockutils [req-eb59d21d-2328-423a-8b52-dc47358f0518 req-12cf86a4-4daa-44b9-ba99-26854c4f4aa8 service nova] Releasing lock "refresh_cache-f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.260348] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.260598] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.260787] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Deleting the datastore file [datastore1] a8b831c7-bd6a-4218-b19e-1c43e3678c59 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.261088] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02372262-ca7a-4894-8cb4-62f057a276d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.269038] env[61936]: DEBUG oslo_vmware.api [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for the task: (returnval){ [ 891.269038] env[61936]: value = "task-1253067" [ 891.269038] env[61936]: _type = "Task" [ 891.269038] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.277981] env[61936]: DEBUG oslo_vmware.api [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.286979] env[61936]: DEBUG nova.network.neutron [-] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.330907] env[61936]: DEBUG nova.network.neutron [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Updating instance_info_cache with network_info: [{"id": "39c4ceec-fef3-4d7d-8a6b-27fc061802a0", "address": "fa:16:3e:be:e8:f6", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39c4ceec-fe", "ovs_interfaceid": "39c4ceec-fef3-4d7d-8a6b-27fc061802a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.391435] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-38c7089d-7085-4ec0-be1d-e8d60ef78cdd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.405263] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84bf2a3-a60f-4044-9371-5045f165e91a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.439730] env[61936]: DEBUG nova.compute.manager [req-70df2aa9-c4de-4a91-9924-8f5228d8d1f3 req-747cbb47-9490-46f1-b65e-2ea131ccb305 service nova] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Detach interface failed, port_id=b7963334-7c72-41c3-9e98-72a837dbb33f, reason: Instance 18175943-35d2-4792-a8ea-e54b4ffbaf2f could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 891.518984] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520a33c7-8005-55f8-581d-66de3632605a, 'name': SearchDatastore_Task, 'duration_secs': 0.009684} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.519746] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb1358fd-31c1-45ff-ad94-b3c8c1693da3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.527997] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 891.527997] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526f26dd-3608-1c96-4378-8a4d8eb3e23a" [ 891.527997] env[61936]: _type = "Task" [ 891.527997] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.540640] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526f26dd-3608-1c96-4378-8a4d8eb3e23a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.559023] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9c9466-c14a-481b-b366-b07f56bd6f38 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.563272] env[61936]: DEBUG nova.compute.manager [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Received event network-vif-plugged-39c4ceec-fef3-4d7d-8a6b-27fc061802a0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 891.563484] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] Acquiring lock "628882ff-2f0f-49ae-8454-d3791fb55545-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.563746] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] Lock "628882ff-2f0f-49ae-8454-d3791fb55545-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.563927] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] Lock "628882ff-2f0f-49ae-8454-d3791fb55545-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.564108] env[61936]: DEBUG nova.compute.manager [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] No waiting events found dispatching network-vif-plugged-39c4ceec-fef3-4d7d-8a6b-27fc061802a0 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 891.564274] env[61936]: WARNING nova.compute.manager [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Received unexpected event network-vif-plugged-39c4ceec-fef3-4d7d-8a6b-27fc061802a0 for instance with vm_state building and task_state spawning. [ 891.564468] env[61936]: DEBUG nova.compute.manager [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Received event network-changed-39c4ceec-fef3-4d7d-8a6b-27fc061802a0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 891.564631] env[61936]: DEBUG nova.compute.manager [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Refreshing instance network info cache due to event network-changed-39c4ceec-fef3-4d7d-8a6b-27fc061802a0. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 891.564802] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] Acquiring lock "refresh_cache-628882ff-2f0f-49ae-8454-d3791fb55545" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.570678] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3660c51-e969-488a-80de-ed2f813003f1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.603335] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299c64e3-e331-4f81-88e2-64f2b450dca6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.612162] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6a800e-3d87-4bf9-84b7-9c8555a1c67a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.628394] env[61936]: DEBUG nova.compute.provider_tree [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.671754] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253065, 'name': Rename_Task, 'duration_secs': 0.171389} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.672053] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.672308] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6015ede-6161-491a-8891-195891fe926f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.680115] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 891.680115] env[61936]: value = "task-1253068" [ 891.680115] env[61936]: _type = "Task" [ 891.680115] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.688125] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253068, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.780390] env[61936]: DEBUG oslo_vmware.api [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Task: {'id': task-1253067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141889} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.780735] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.780968] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 891.781302] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 891.781524] env[61936]: INFO nova.compute.manager [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Took 1.15 seconds to destroy the instance on the hypervisor. [ 891.781774] env[61936]: DEBUG oslo.service.loopingcall [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.781955] env[61936]: DEBUG nova.compute.manager [-] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 891.782108] env[61936]: DEBUG nova.network.neutron [-] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 891.789217] env[61936]: INFO nova.compute.manager [-] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Took 1.77 seconds to deallocate network for instance. [ 891.835042] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-628882ff-2f0f-49ae-8454-d3791fb55545" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.835383] env[61936]: DEBUG nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Instance network_info: |[{"id": "39c4ceec-fef3-4d7d-8a6b-27fc061802a0", "address": "fa:16:3e:be:e8:f6", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39c4ceec-fe", "ovs_interfaceid": "39c4ceec-fef3-4d7d-8a6b-27fc061802a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 891.835687] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] Acquired lock "refresh_cache-628882ff-2f0f-49ae-8454-d3791fb55545" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.835871] env[61936]: DEBUG nova.network.neutron [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Refreshing network info cache for port 39c4ceec-fef3-4d7d-8a6b-27fc061802a0 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.837690] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:e8:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '39c4ceec-fef3-4d7d-8a6b-27fc061802a0', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.846017] env[61936]: DEBUG oslo.service.loopingcall [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.846907] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.847166] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9165de4d-69c6-4a98-a677-a7a21aea0767 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.869268] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.869268] env[61936]: value = "task-1253069" [ 891.869268] env[61936]: _type = "Task" [ 891.869268] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.883814] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253069, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.047704] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526f26dd-3608-1c96-4378-8a4d8eb3e23a, 'name': SearchDatastore_Task, 'duration_secs': 0.012772} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.048048] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.048382] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] f71581c9-b754-4ee4-bf47-ea8e52e4a2f6/f71581c9-b754-4ee4-bf47-ea8e52e4a2f6.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.048685] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b3c2e3fa-5869-4ef1-81a1-ac1153590adf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.057329] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 892.057329] env[61936]: value = "task-1253070" [ 892.057329] env[61936]: _type = "Task" [ 892.057329] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.067090] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.132634] env[61936]: DEBUG nova.scheduler.client.report [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 892.204852] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253068, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.213933] env[61936]: DEBUG nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 892.223016] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 892.223544] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 892.243050] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 892.243331] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.243546] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 892.243766] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.243920] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 892.244080] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 892.244297] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 892.244470] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 892.244642] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 892.244817] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 892.245014] env[61936]: DEBUG nova.virt.hardware [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 892.246214] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6bce17-d884-4920-9c81-ec312a9eb0e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.256267] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d48763-6390-4900-9bac-f121fe3ab403 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.295756] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.382375] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253069, 'name': CreateVM_Task, 'duration_secs': 0.382354} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.382550] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.383295] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.383467] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.383851] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.384145] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f2ef743-989b-4459-8c70-73cd7e0b128d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.391489] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 892.391489] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5208ff37-2a53-292a-cf86-4be06054117f" [ 892.391489] env[61936]: _type = "Task" [ 892.391489] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.401392] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5208ff37-2a53-292a-cf86-4be06054117f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.545073] env[61936]: DEBUG nova.network.neutron [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Successfully updated port: 691101d8-d631-42c1-80ce-2aba05214046 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.568302] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253070, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505918} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.568615] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] f71581c9-b754-4ee4-bf47-ea8e52e4a2f6/f71581c9-b754-4ee4-bf47-ea8e52e4a2f6.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.569361] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.569361] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0017d0c-ac87-4053-a594-2edf122e6a47 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.576031] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 892.576031] env[61936]: value = "task-1253071" [ 892.576031] env[61936]: _type = "Task" [ 892.576031] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.584647] env[61936]: DEBUG nova.network.neutron [-] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.586044] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253071, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.601998] env[61936]: DEBUG nova.network.neutron [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Updated VIF entry in instance network info cache for port 39c4ceec-fef3-4d7d-8a6b-27fc061802a0. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.602395] env[61936]: DEBUG nova.network.neutron [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Updating instance_info_cache with network_info: [{"id": "39c4ceec-fef3-4d7d-8a6b-27fc061802a0", "address": "fa:16:3e:be:e8:f6", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39c4ceec-fe", "ovs_interfaceid": "39c4ceec-fef3-4d7d-8a6b-27fc061802a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.641385] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.642157] env[61936]: DEBUG nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 892.647570] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.701s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.649412] env[61936]: INFO nova.compute.claims [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.697861] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253068, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.732792] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 892.732792] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 892.732792] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Rebuilding the list of instances to heal {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10330}} [ 892.903052] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5208ff37-2a53-292a-cf86-4be06054117f, 'name': SearchDatastore_Task, 'duration_secs': 0.047698} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.903362] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.903602] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.903833] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.903977] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.904169] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.904431] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7858947d-fd48-448b-a340-ca86b857ba70 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.913524] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.913710] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 892.914428] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4b3f1d3-0c5b-457f-a621-09fc74a9c304 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.919946] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 892.919946] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f60030-79fb-a8ea-86e0-c79055207ea9" [ 892.919946] env[61936]: _type = "Task" [ 892.919946] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.931307] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f60030-79fb-a8ea-86e0-c79055207ea9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.954997] env[61936]: DEBUG nova.compute.manager [req-be141b46-6aa8-4315-a580-146819f4d01e req-f5bfd50b-3ea5-40a0-a2d3-dd8433a6b8f9 service nova] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Received event network-vif-deleted-14553288-1c58-403e-bd50-95de5e72fed1 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 893.046724] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-9f6a1c42-e5a5-4f1d-9740-35dc6028183a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.046883] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-9f6a1c42-e5a5-4f1d-9740-35dc6028183a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.047029] env[61936]: DEBUG nova.network.neutron [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.085603] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253071, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067273} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.085881] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.086712] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6ab187-fd6d-4564-9bf9-3598d5ff5faf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.089320] env[61936]: INFO nova.compute.manager [-] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Took 1.31 seconds to deallocate network for instance. [ 893.113224] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] f71581c9-b754-4ee4-bf47-ea8e52e4a2f6/f71581c9-b754-4ee4-bf47-ea8e52e4a2f6.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.113837] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f37b2f9-cf4d-4e83-a438-c122a647e738 req-1b5bae0a-8cd7-4fee-b771-cc30a4c2f152 service nova] Releasing lock "refresh_cache-628882ff-2f0f-49ae-8454-d3791fb55545" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.114191] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52d09392-5bb1-4e07-9e99-12aadb13f725 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.136315] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 893.136315] env[61936]: value = "task-1253072" [ 893.136315] env[61936]: _type = "Task" [ 893.136315] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.144948] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253072, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.155806] env[61936]: DEBUG nova.compute.utils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.160923] env[61936]: DEBUG nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 893.162110] env[61936]: DEBUG nova.network.neutron [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 893.192240] env[61936]: DEBUG oslo_vmware.api [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253068, 'name': PowerOnVM_Task, 'duration_secs': 1.16397} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.192562] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.192738] env[61936]: INFO nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Took 9.43 seconds to spawn the instance on the hypervisor. [ 893.192916] env[61936]: DEBUG nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 893.193724] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62d85a3-e9ab-4b96-a32f-79358fb8b40d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.203574] env[61936]: DEBUG nova.policy [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9260b79bb7f4e44bb1f50b16273df09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ce9d789b30c49758bd073d17b3f1281', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.241252] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Skipping network cache update for instance because it is being deleted. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 893.241431] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 893.241551] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 893.241663] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 893.241779] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 893.241921] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 893.242062] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Skipping network cache update for instance because it is Building. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10339}} [ 893.269740] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.269888] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquired lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.270031] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Forcefully refreshing network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 893.270193] env[61936]: DEBUG nova.objects.instance [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lazy-loading 'info_cache' on Instance uuid 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.430723] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f60030-79fb-a8ea-86e0-c79055207ea9, 'name': SearchDatastore_Task, 'duration_secs': 0.009852} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.431543] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79086853-6b9c-49a6-806f-4db445868533 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.437241] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 893.437241] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524455c1-3128-258c-28ed-c56f18dc7e11" [ 893.437241] env[61936]: _type = "Task" [ 893.437241] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.445017] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524455c1-3128-258c-28ed-c56f18dc7e11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.587454] env[61936]: DEBUG nova.network.neutron [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 893.595048] env[61936]: DEBUG nova.compute.manager [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Received event network-vif-plugged-691101d8-d631-42c1-80ce-2aba05214046 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 893.595284] env[61936]: DEBUG oslo_concurrency.lockutils [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] Acquiring lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.595494] env[61936]: DEBUG oslo_concurrency.lockutils [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.595663] env[61936]: DEBUG oslo_concurrency.lockutils [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.595831] env[61936]: DEBUG nova.compute.manager [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] No waiting events found dispatching network-vif-plugged-691101d8-d631-42c1-80ce-2aba05214046 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 893.595994] env[61936]: WARNING nova.compute.manager [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Received unexpected event network-vif-plugged-691101d8-d631-42c1-80ce-2aba05214046 for instance with vm_state building and task_state spawning. [ 893.596582] env[61936]: DEBUG nova.compute.manager [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Received event network-changed-691101d8-d631-42c1-80ce-2aba05214046 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 893.596761] env[61936]: DEBUG nova.compute.manager [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Refreshing instance network info cache due to event network-changed-691101d8-d631-42c1-80ce-2aba05214046. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 893.596935] env[61936]: DEBUG oslo_concurrency.lockutils [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] Acquiring lock "refresh_cache-9f6a1c42-e5a5-4f1d-9740-35dc6028183a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.599061] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.639437] env[61936]: DEBUG nova.network.neutron [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Successfully created port: 894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.651385] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253072, 'name': ReconfigVM_Task, 'duration_secs': 0.291852} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.651667] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Reconfigured VM instance instance-00000048 to attach disk [datastore2] f71581c9-b754-4ee4-bf47-ea8e52e4a2f6/f71581c9-b754-4ee4-bf47-ea8e52e4a2f6.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.652322] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7478ba5-10a8-4708-9ff4-8b089ac0cabc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.658903] env[61936]: DEBUG nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 893.665379] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 893.665379] env[61936]: value = "task-1253073" [ 893.665379] env[61936]: _type = "Task" [ 893.665379] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.677653] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253073, 'name': Rename_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.719714] env[61936]: INFO nova.compute.manager [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Took 33.93 seconds to build instance. [ 893.768246] env[61936]: DEBUG nova.network.neutron [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Updating instance_info_cache with network_info: [{"id": "691101d8-d631-42c1-80ce-2aba05214046", "address": "fa:16:3e:22:54:e6", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap691101d8-d6", "ovs_interfaceid": "691101d8-d631-42c1-80ce-2aba05214046", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.950333] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524455c1-3128-258c-28ed-c56f18dc7e11, 'name': SearchDatastore_Task, 'duration_secs': 0.00939} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.950605] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.950974] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 628882ff-2f0f-49ae-8454-d3791fb55545/628882ff-2f0f-49ae-8454-d3791fb55545.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.951117] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbcf6604-409d-473b-b857-09e3438a5c80 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.959396] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 893.959396] env[61936]: value = "task-1253074" [ 893.959396] env[61936]: _type = "Task" [ 893.959396] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.968729] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253074, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.023374] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db1fbb2-663d-4bf1-bf4d-e8cc55af9128 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.032138] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3939c3cd-a855-4edf-9df5-e55b748897f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.063352] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cdadee3-b0e9-49ba-bfef-77ba38ae0ab6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.072049] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414c17f6-4b16-4c66-8040-1cdb38b18384 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.085906] env[61936]: DEBUG nova.compute.provider_tree [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.183473] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253073, 'name': Rename_Task, 'duration_secs': 0.162968} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.183746] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.184055] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-124b9192-bbf7-4cf2-bc0b-78271c417dab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.195491] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 894.195491] env[61936]: value = "task-1253075" [ 894.195491] env[61936]: _type = "Task" [ 894.195491] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.204787] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253075, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.222420] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ba904057-2ac7-4940-95ad-fcda811bdad4 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.348s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.271152] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-9f6a1c42-e5a5-4f1d-9740-35dc6028183a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.271797] env[61936]: DEBUG nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Instance network_info: |[{"id": "691101d8-d631-42c1-80ce-2aba05214046", "address": "fa:16:3e:22:54:e6", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap691101d8-d6", "ovs_interfaceid": "691101d8-d631-42c1-80ce-2aba05214046", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 894.272317] env[61936]: DEBUG oslo_concurrency.lockutils [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] Acquired lock "refresh_cache-9f6a1c42-e5a5-4f1d-9740-35dc6028183a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.272904] env[61936]: DEBUG nova.network.neutron [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Refreshing network info cache for port 691101d8-d631-42c1-80ce-2aba05214046 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 894.276489] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:54:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '691101d8-d631-42c1-80ce-2aba05214046', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.284472] env[61936]: DEBUG oslo.service.loopingcall [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.288145] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.288938] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4df3dbe-5fb9-420f-b797-128f963cee75 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.315109] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.315109] env[61936]: value = "task-1253076" [ 894.315109] env[61936]: _type = "Task" [ 894.315109] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.327829] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253076, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.470514] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253074, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494385} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.470966] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 628882ff-2f0f-49ae-8454-d3791fb55545/628882ff-2f0f-49ae-8454-d3791fb55545.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 894.471191] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.471545] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7ca4a3a-af47-4818-b7f8-7a7fe637afc7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.479154] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 894.479154] env[61936]: value = "task-1253077" [ 894.479154] env[61936]: _type = "Task" [ 894.479154] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.489981] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.588458] env[61936]: DEBUG nova.scheduler.client.report [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 894.674605] env[61936]: DEBUG nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 894.701668] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 894.701964] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.702218] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 894.702445] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.702549] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 894.702703] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 894.702914] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 894.703628] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 894.703628] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 894.703769] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 894.703962] env[61936]: DEBUG nova.virt.hardware [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 894.705316] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8a6535-020c-4931-896c-498657ddf2f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.711632] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253075, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.717125] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6591982f-e8bd-4e8c-a859-6772d4b7ea51 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.732263] env[61936]: DEBUG nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 894.826016] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253076, 'name': CreateVM_Task, 'duration_secs': 0.431391} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.826016] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.826503] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.826671] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.826996] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.827274] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3a7f80f-079f-4717-9cdd-8c06bf55b0ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.832903] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 894.832903] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525d99f7-d171-0baa-1718-af61303e9c83" [ 894.832903] env[61936]: _type = "Task" [ 894.832903] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.841334] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525d99f7-d171-0baa-1718-af61303e9c83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.952377] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.952722] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.952940] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.953138] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.953339] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.957481] env[61936]: INFO nova.compute.manager [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Terminating instance [ 894.989374] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063861} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.989658] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.990677] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6696ec-8b4f-40a6-968c-36de17d1cd29 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.014676] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 628882ff-2f0f-49ae-8454-d3791fb55545/628882ff-2f0f-49ae-8454-d3791fb55545.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.016516] env[61936]: DEBUG nova.network.neutron [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Updated VIF entry in instance network info cache for port 691101d8-d631-42c1-80ce-2aba05214046. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.019549] env[61936]: DEBUG nova.network.neutron [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Updating instance_info_cache with network_info: [{"id": "691101d8-d631-42c1-80ce-2aba05214046", "address": "fa:16:3e:22:54:e6", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap691101d8-d6", "ovs_interfaceid": "691101d8-d631-42c1-80ce-2aba05214046", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.019549] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83ddddf3-6586-4244-93fd-ffa4ed9761d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.047826] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 895.047826] env[61936]: value = "task-1253078" [ 895.047826] env[61936]: _type = "Task" [ 895.047826] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.057197] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253078, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.086200] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updating instance_info_cache with network_info: [{"id": "a342f5d6-5a18-4e6f-b046-fd7365684711", "address": "fa:16:3e:93:27:c5", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa342f5d6-5a", "ovs_interfaceid": "a342f5d6-5a18-4e6f-b046-fd7365684711", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.095577] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.095890] env[61936]: DEBUG nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 895.098761] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.652s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.099059] env[61936]: DEBUG nova.objects.instance [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 895.140291] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-e15be950-1c32-4801-b4e4-a4d31ff0cfec-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.140503] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-e15be950-1c32-4801-b4e4-a4d31ff0cfec-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.140861] env[61936]: DEBUG nova.objects.instance [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'flavor' on Instance uuid e15be950-1c32-4801-b4e4-a4d31ff0cfec {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.205484] env[61936]: DEBUG oslo_vmware.api [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253075, 'name': PowerOnVM_Task, 'duration_secs': 0.644254} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.205782] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.205984] env[61936]: INFO nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Took 7.93 seconds to spawn the instance on the hypervisor. [ 895.206184] env[61936]: DEBUG nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 895.206956] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437f84e8-35fe-4f05-9f7b-2ec554a4298e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.255625] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.345566] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525d99f7-d171-0baa-1718-af61303e9c83, 'name': SearchDatastore_Task, 'duration_secs': 0.079511} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.345785] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.346028] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.346311] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.346531] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.346735] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.346996] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4d9ea5db-5059-43a4-a73e-a7bbd5c74135 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.356624] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.356818] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.357601] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe06d2b5-f284-45e4-85b3-e3754a37f712 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.363228] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 895.363228] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5244b4bd-0034-9df1-9000-354743ae0043" [ 895.363228] env[61936]: _type = "Task" [ 895.363228] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.375016] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5244b4bd-0034-9df1-9000-354743ae0043, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.430158] env[61936]: DEBUG nova.network.neutron [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Successfully updated port: 894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.462390] env[61936]: DEBUG nova.compute.manager [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 895.462669] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.463590] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50398edd-8218-4148-8db9-305dfdf2a115 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.473270] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.473928] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f91bb0f-6b4b-498e-9179-7b9ebd7a9997 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.483651] env[61936]: DEBUG oslo_vmware.api [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 895.483651] env[61936]: value = "task-1253079" [ 895.483651] env[61936]: _type = "Task" [ 895.483651] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.493388] env[61936]: DEBUG oslo_vmware.api [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253079, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.540144] env[61936]: DEBUG oslo_concurrency.lockutils [req-0c9b9d13-d926-4f61-8ccd-4f7bbf579a90 req-f5a98a67-4911-49b7-87cf-1d85cecd854b service nova] Releasing lock "refresh_cache-9f6a1c42-e5a5-4f1d-9740-35dc6028183a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.558752] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.589186] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Releasing lock "refresh_cache-7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.589345] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updated the network info_cache for instance {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10397}} [ 895.589582] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.589767] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.589914] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.590098] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.590298] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.591073] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.591073] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 895.591073] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 895.603736] env[61936]: DEBUG nova.compute.utils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 895.609281] env[61936]: DEBUG nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 895.609281] env[61936]: DEBUG nova.network.neutron [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 895.658199] env[61936]: DEBUG nova.policy [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ab09da635d44a9992809c35449ba07d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c21a56ef57b4d78bb6b1e6343efc442', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 895.725563] env[61936]: INFO nova.compute.manager [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Took 33.03 seconds to build instance. [ 895.764346] env[61936]: DEBUG nova.objects.instance [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'pci_requests' on Instance uuid e15be950-1c32-4801-b4e4-a4d31ff0cfec {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.857898] env[61936]: DEBUG nova.compute.manager [req-137ea997-59ab-402d-9d1b-d2e2805f1713 req-416d0ee2-451a-42fd-ad01-ecaf7d607e7f service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Received event network-vif-plugged-894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 895.858148] env[61936]: DEBUG oslo_concurrency.lockutils [req-137ea997-59ab-402d-9d1b-d2e2805f1713 req-416d0ee2-451a-42fd-ad01-ecaf7d607e7f service nova] Acquiring lock "10669072-a8e8-4262-b390-8bf31af6fafe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.858416] env[61936]: DEBUG oslo_concurrency.lockutils [req-137ea997-59ab-402d-9d1b-d2e2805f1713 req-416d0ee2-451a-42fd-ad01-ecaf7d607e7f service nova] Lock "10669072-a8e8-4262-b390-8bf31af6fafe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.858623] env[61936]: DEBUG oslo_concurrency.lockutils [req-137ea997-59ab-402d-9d1b-d2e2805f1713 req-416d0ee2-451a-42fd-ad01-ecaf7d607e7f service nova] Lock "10669072-a8e8-4262-b390-8bf31af6fafe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.858739] env[61936]: DEBUG nova.compute.manager [req-137ea997-59ab-402d-9d1b-d2e2805f1713 req-416d0ee2-451a-42fd-ad01-ecaf7d607e7f service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] No waiting events found dispatching network-vif-plugged-894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 895.858908] env[61936]: WARNING nova.compute.manager [req-137ea997-59ab-402d-9d1b-d2e2805f1713 req-416d0ee2-451a-42fd-ad01-ecaf7d607e7f service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Received unexpected event network-vif-plugged-894b1eac-a089-4cf8-adbc-f2e646971faf for instance with vm_state building and task_state spawning. [ 895.873971] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5244b4bd-0034-9df1-9000-354743ae0043, 'name': SearchDatastore_Task, 'duration_secs': 0.016582} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.874772] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fa149ce-376f-41be-80e1-0beebbafea8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.880504] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 895.880504] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239bd97-5be3-1557-d57b-ff018eec4b50" [ 895.880504] env[61936]: _type = "Task" [ 895.880504] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.888492] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239bd97-5be3-1557-d57b-ff018eec4b50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.933993] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.933993] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.933993] env[61936]: DEBUG nova.network.neutron [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.994259] env[61936]: DEBUG oslo_vmware.api [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253079, 'name': PowerOffVM_Task, 'duration_secs': 0.195666} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.994532] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.994701] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.994957] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e44fdab-be34-45b4-9222-7c96818dd1d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.035873] env[61936]: DEBUG nova.network.neutron [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Successfully created port: d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.058538] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253078, 'name': ReconfigVM_Task, 'duration_secs': 0.612312} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.058694] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 628882ff-2f0f-49ae-8454-d3791fb55545/628882ff-2f0f-49ae-8454-d3791fb55545.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.059389] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b3ba4b4-5df1-4beb-9f58-a4ef7d3b4ffe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.067109] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 896.067109] env[61936]: value = "task-1253081" [ 896.067109] env[61936]: _type = "Task" [ 896.067109] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.076045] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253081, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.080240] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 896.080240] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 896.080240] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Deleting the datastore file [datastore1] 3af90345-f5bf-4d25-90ac-c5389aef5c8e {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 896.080547] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55b35f75-560f-4975-808e-9e65d17e5af0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.087730] env[61936]: DEBUG oslo_vmware.api [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for the task: (returnval){ [ 896.087730] env[61936]: value = "task-1253082" [ 896.087730] env[61936]: _type = "Task" [ 896.087730] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.094032] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.099113] env[61936]: DEBUG oslo_vmware.api [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253082, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.109729] env[61936]: DEBUG nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 896.113208] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6b190913-d5fa-4586-a767-55f7234ec235 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.114333] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.870s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.115836] env[61936]: INFO nova.compute.claims [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.227758] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ccf38cb-1647-4340-904a-10546390b0cb tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.844s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.267018] env[61936]: DEBUG nova.objects.base [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 896.267297] env[61936]: DEBUG nova.network.neutron [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.395361] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239bd97-5be3-1557-d57b-ff018eec4b50, 'name': SearchDatastore_Task, 'duration_secs': 0.022323} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.395361] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.395361] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 9f6a1c42-e5a5-4f1d-9740-35dc6028183a/9f6a1c42-e5a5-4f1d-9740-35dc6028183a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.395361] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5942b100-b9af-49db-a1c0-f87d2fcc7893 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.405021] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 896.405021] env[61936]: value = "task-1253083" [ 896.405021] env[61936]: _type = "Task" [ 896.405021] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.412345] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.471182] env[61936]: DEBUG nova.network.neutron [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 896.551352] env[61936]: DEBUG nova.policy [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.578398] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253081, 'name': Rename_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.602029] env[61936]: DEBUG oslo_vmware.api [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Task: {'id': task-1253082, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388363} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.602485] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.602999] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.605601] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.605601] env[61936]: INFO nova.compute.manager [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 896.605601] env[61936]: DEBUG oslo.service.loopingcall [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.605601] env[61936]: DEBUG nova.compute.manager [-] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 896.605601] env[61936]: DEBUG nova.network.neutron [-] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 896.643656] env[61936]: DEBUG nova.network.neutron [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Updating instance_info_cache with network_info: [{"id": "894b1eac-a089-4cf8-adbc-f2e646971faf", "address": "fa:16:3e:f9:eb:61", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894b1eac-a0", "ovs_interfaceid": "894b1eac-a089-4cf8-adbc-f2e646971faf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.913190] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.079337] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253081, 'name': Rename_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.120584] env[61936]: DEBUG nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 897.132526] env[61936]: DEBUG nova.network.neutron [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Successfully created port: 7d859e6c-be30-4b94-9fb6-6af6858a2ba5 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.146483] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.146902] env[61936]: DEBUG nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Instance network_info: |[{"id": "894b1eac-a089-4cf8-adbc-f2e646971faf", "address": "fa:16:3e:f9:eb:61", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894b1eac-a0", "ovs_interfaceid": "894b1eac-a089-4cf8-adbc-f2e646971faf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 897.148959] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 897.149210] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.149419] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 897.149619] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.149769] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 897.149917] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 897.150147] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 897.150310] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 897.150479] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 897.150642] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 897.150811] env[61936]: DEBUG nova.virt.hardware [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 897.151223] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:eb:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '894b1eac-a089-4cf8-adbc-f2e646971faf', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.159688] env[61936]: DEBUG oslo.service.loopingcall [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.160398] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7865bda7-3afa-4721-b1bd-acde8f4bd7b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.166335] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.166732] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cfbb91bc-e0db-4626-b2b3-b57c198a816e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.192917] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05b20f0-0596-4774-92f6-2e4defaa166b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.201032] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.201032] env[61936]: value = "task-1253084" [ 897.201032] env[61936]: _type = "Task" [ 897.201032] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.222395] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253084, 'name': CreateVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.375724] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.375997] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.376253] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.376413] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.376575] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.378812] env[61936]: INFO nova.compute.manager [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Terminating instance [ 897.414393] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253083, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.509156] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36bbe24-a8f8-44b7-b5ab-dfe9d56768b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.519264] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7132f17-e3e4-4231-a6a4-b5cb739cdc02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.555318] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8a34ca-1774-4f75-b92e-5ea7687870ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.558261] env[61936]: DEBUG nova.network.neutron [-] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.565911] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6aa1971-af10-40c2-b6e4-a021b2cf7896 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.580189] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253081, 'name': Rename_Task, 'duration_secs': 1.196099} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.587830] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.588781] env[61936]: DEBUG nova.compute.provider_tree [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.589997] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b433c31-a81a-4b1a-a2f2-d78dcc5b562a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.598886] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 897.598886] env[61936]: value = "task-1253085" [ 897.598886] env[61936]: _type = "Task" [ 897.598886] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.611458] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253085, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.721556] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253084, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.885278] env[61936]: DEBUG nova.compute.manager [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 897.885533] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.886537] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb484c12-aac2-4fa5-9e96-1e2831c52c0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.894513] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.894762] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5af538f0-9198-4a19-8e2b-4f07f1948c9f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.901571] env[61936]: DEBUG oslo_vmware.api [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 897.901571] env[61936]: value = "task-1253086" [ 897.901571] env[61936]: _type = "Task" [ 897.901571] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.914425] env[61936]: DEBUG oslo_vmware.api [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.917890] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253083, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.293692} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.918203] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 9f6a1c42-e5a5-4f1d-9740-35dc6028183a/9f6a1c42-e5a5-4f1d-9740-35dc6028183a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.918518] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.918843] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d7ab26c-c448-4e60-a5e0-f71db4b2506f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.926789] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 897.926789] env[61936]: value = "task-1253087" [ 897.926789] env[61936]: _type = "Task" [ 897.926789] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.935716] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253087, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.061204] env[61936]: INFO nova.compute.manager [-] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Took 1.46 seconds to deallocate network for instance. [ 898.093814] env[61936]: DEBUG nova.scheduler.client.report [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 898.119145] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253085, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.185414] env[61936]: DEBUG nova.network.neutron [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Successfully updated port: d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.221682] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253084, 'name': CreateVM_Task, 'duration_secs': 0.639243} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.221871] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.222578] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.222750] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.223091] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.223415] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e61db93-49e6-4763-9f1c-b55bdd6a23f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.228942] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 898.228942] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52630e5c-9890-1bd4-ab13-ada7405db7cf" [ 898.228942] env[61936]: _type = "Task" [ 898.228942] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.236972] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52630e5c-9890-1bd4-ab13-ada7405db7cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.388339] env[61936]: DEBUG nova.compute.manager [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Received event network-changed-894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 898.388551] env[61936]: DEBUG nova.compute.manager [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Refreshing instance network info cache due to event network-changed-894b1eac-a089-4cf8-adbc-f2e646971faf. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 898.388760] env[61936]: DEBUG oslo_concurrency.lockutils [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] Acquiring lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.388902] env[61936]: DEBUG oslo_concurrency.lockutils [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] Acquired lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.389077] env[61936]: DEBUG nova.network.neutron [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Refreshing network info cache for port 894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.413153] env[61936]: DEBUG oslo_vmware.api [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253086, 'name': PowerOffVM_Task, 'duration_secs': 0.188024} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.413404] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 898.413604] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.413858] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e2e0ccc-4d56-45c7-97e4-d3c676db6276 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.435469] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253087, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082182} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.435743] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.436513] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2970dbd5-1817-4b02-bb2f-73b05a168fee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.459136] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 9f6a1c42-e5a5-4f1d-9740-35dc6028183a/9f6a1c42-e5a5-4f1d-9740-35dc6028183a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.459777] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9769ce5a-7edc-486c-a9aa-ec6cab78cb91 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.481744] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 898.481744] env[61936]: value = "task-1253089" [ 898.481744] env[61936]: _type = "Task" [ 898.481744] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.497406] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253089, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.520978] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.521137] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.521329] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleting the datastore file [datastore2] f71581c9-b754-4ee4-bf47-ea8e52e4a2f6 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.521609] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b494abd6-a7e1-4d34-afcf-2a5cce4acb1a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.529924] env[61936]: DEBUG oslo_vmware.api [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 898.529924] env[61936]: value = "task-1253090" [ 898.529924] env[61936]: _type = "Task" [ 898.529924] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.538299] env[61936]: DEBUG oslo_vmware.api [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253090, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.569152] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.599564] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.599968] env[61936]: DEBUG nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 898.602717] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.695s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.602950] env[61936]: DEBUG nova.objects.instance [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lazy-loading 'resources' on Instance uuid e800f5e9-deb0-4592-aded-3768d238e311 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.613844] env[61936]: DEBUG oslo_vmware.api [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253085, 'name': PowerOnVM_Task, 'duration_secs': 0.558132} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.614782] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.615053] env[61936]: INFO nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Took 8.89 seconds to spawn the instance on the hypervisor. [ 898.615284] env[61936]: DEBUG nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 898.616192] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d77544a-36c0-4e88-aab8-364511b3c494 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.687554] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.687740] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.687899] env[61936]: DEBUG nova.network.neutron [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.741698] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52630e5c-9890-1bd4-ab13-ada7405db7cf, 'name': SearchDatastore_Task, 'duration_secs': 0.02274} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.741698] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.741698] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.741698] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.741984] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.741984] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.742283] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4bdf85b-bda4-4ba9-9028-85470047a805 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.752874] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.753092] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.753952] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d99b14c-cdc1-4b4f-9257-529a1326f5aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.759491] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 898.759491] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dc2209-db41-bc0c-29f2-70b9b7300fd8" [ 898.759491] env[61936]: _type = "Task" [ 898.759491] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.767805] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dc2209-db41-bc0c-29f2-70b9b7300fd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.961483] env[61936]: DEBUG nova.network.neutron [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Successfully updated port: 7d859e6c-be30-4b94-9fb6-6af6858a2ba5 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.998890] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253089, 'name': ReconfigVM_Task, 'duration_secs': 0.357501} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.998890] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 9f6a1c42-e5a5-4f1d-9740-35dc6028183a/9f6a1c42-e5a5-4f1d-9740-35dc6028183a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.999118] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-55ad51fb-5ad9-4236-aecc-39f5d8fe9aad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.010986] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 899.010986] env[61936]: value = "task-1253091" [ 899.010986] env[61936]: _type = "Task" [ 899.010986] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.020604] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253091, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.039716] env[61936]: DEBUG oslo_vmware.api [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253090, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.105616] env[61936]: DEBUG nova.compute.utils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 899.107010] env[61936]: DEBUG nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 899.107173] env[61936]: DEBUG nova.network.neutron [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.136195] env[61936]: INFO nova.compute.manager [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Took 35.16 seconds to build instance. [ 899.229966] env[61936]: DEBUG nova.policy [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '452043da648742e28ece5659906fc6a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2171cb70a814ecb986099f6149ffc89', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.267092] env[61936]: DEBUG nova.network.neutron [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 899.275011] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dc2209-db41-bc0c-29f2-70b9b7300fd8, 'name': SearchDatastore_Task, 'duration_secs': 0.03756} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.276088] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7f064fd-fdc1-47c8-a9e7-34a7d8a0bae6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.282877] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 899.282877] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e6e863-df44-39fd-560e-3fa51d412815" [ 899.282877] env[61936]: _type = "Task" [ 899.282877] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.292152] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e6e863-df44-39fd-560e-3fa51d412815, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.326496] env[61936]: DEBUG nova.network.neutron [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Updated VIF entry in instance network info cache for port 894b1eac-a089-4cf8-adbc-f2e646971faf. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.327087] env[61936]: DEBUG nova.network.neutron [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Updating instance_info_cache with network_info: [{"id": "894b1eac-a089-4cf8-adbc-f2e646971faf", "address": "fa:16:3e:f9:eb:61", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894b1eac-a0", "ovs_interfaceid": "894b1eac-a089-4cf8-adbc-f2e646971faf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.463936] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.464139] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.464322] env[61936]: DEBUG nova.network.neutron [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.468410] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63bdbc8f-8328-4c19-a1f4-a05712dc01b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.476855] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4faa7551-461e-409d-aa4e-84ddb5602a0e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.511191] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5e7413-ba3f-4fb6-aa62-cf6309d5518c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.523075] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253091, 'name': Rename_Task, 'duration_secs': 0.151885} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.527618] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.528331] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07794e83-8ef1-4c67-a3d4-2070503cd201 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.531056] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a456274-53c9-4e60-96c8-367a29710c61 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.541185] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 899.541185] env[61936]: value = "task-1253092" [ 899.541185] env[61936]: _type = "Task" [ 899.541185] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.552744] env[61936]: DEBUG nova.compute.provider_tree [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.553914] env[61936]: DEBUG oslo_vmware.api [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253090, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.547123} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.557266] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.557540] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.557838] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.558148] env[61936]: INFO nova.compute.manager [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Took 1.67 seconds to destroy the instance on the hypervisor. [ 899.558496] env[61936]: DEBUG oslo.service.loopingcall [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.558807] env[61936]: DEBUG nova.compute.manager [-] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 899.558965] env[61936]: DEBUG nova.network.neutron [-] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 899.568976] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253092, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.610423] env[61936]: DEBUG nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 899.625130] env[61936]: DEBUG nova.network.neutron [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updating instance_info_cache with network_info: [{"id": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "address": "fa:16:3e:bb:c3:ef", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0bdd1e2-85", "ovs_interfaceid": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.638538] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0750e412-c96b-4adf-82c9-4ef2ad438ae1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.218s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.794603] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e6e863-df44-39fd-560e-3fa51d412815, 'name': SearchDatastore_Task, 'duration_secs': 0.03807} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.794907] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.795195] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 10669072-a8e8-4262-b390-8bf31af6fafe/10669072-a8e8-4262-b390-8bf31af6fafe.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.795490] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4976600-57b0-4e5a-b0c7-a2b32a33d61d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.804207] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 899.804207] env[61936]: value = "task-1253093" [ 899.804207] env[61936]: _type = "Task" [ 899.804207] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.813812] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.829790] env[61936]: DEBUG oslo_concurrency.lockutils [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] Releasing lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.830083] env[61936]: DEBUG nova.compute.manager [req-8c5511c1-bd32-4716-92ff-cff3d4d7b8f2 req-9013470e-e1f1-47c7-a701-9cb921192141 service nova] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Received event network-vif-deleted-464d9afb-54f5-47c5-aedf-ac53cb44d6b9 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 899.957842] env[61936]: DEBUG nova.network.neutron [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Successfully created port: db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.057507] env[61936]: DEBUG nova.scheduler.client.report [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 900.070466] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253092, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.082754] env[61936]: WARNING nova.network.neutron [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] 03166185-d5e0-477b-892e-20bea284c8db already exists in list: networks containing: ['03166185-d5e0-477b-892e-20bea284c8db']. ignoring it [ 900.115306] env[61936]: INFO nova.virt.block_device [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Booting with volume 612c642f-abe9-4886-ac2f-baadfe0dcb78 at /dev/sda [ 900.127740] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Releasing lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.128329] env[61936]: DEBUG nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Instance network_info: |[{"id": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "address": "fa:16:3e:bb:c3:ef", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0bdd1e2-85", "ovs_interfaceid": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 900.128624] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:c3:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b033f4d-2e92-4702-add6-410a29d3f251', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.136950] env[61936]: DEBUG oslo.service.loopingcall [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.137301] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.137565] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df9db709-c39c-43b0-87fb-2d70e749cb71 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.162603] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.162603] env[61936]: value = "task-1253094" [ 900.162603] env[61936]: _type = "Task" [ 900.162603] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.171031] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-931f4b79-fcbc-4f0d-8947-14831ccd1666 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.179055] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253094, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.189819] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6e6a8b-dc1a-462c-95ab-3e9ee3b5e061 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.223647] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16f59525-fd8f-4604-b74b-a6b77c4a2473 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.234285] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce3320c-fff1-4a75-86ef-bb9b8cd0e6a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.276402] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3c9783-f910-4915-aa63-ad8f61d6d769 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.285579] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d68ee0-90e7-44fa-a255-ae948706195b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.310135] env[61936]: DEBUG nova.virt.block_device [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updating existing volume attachment record: 65da5d25-fafb-4fe8-98d2-19c7a0097487 {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 900.324152] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253093, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.568110] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.963s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.568110] env[61936]: DEBUG oslo_vmware.api [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253092, 'name': PowerOnVM_Task, 'duration_secs': 0.84547} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.568299] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.574s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.570517] env[61936]: INFO nova.compute.claims [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.574837] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.575640] env[61936]: INFO nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Took 8.36 seconds to spawn the instance on the hypervisor. [ 900.575977] env[61936]: DEBUG nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 900.576688] env[61936]: DEBUG nova.network.neutron [-] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.579232] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e784b4-f1c4-4722-9f19-741ca7870e17 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.614351] env[61936]: INFO nova.scheduler.client.report [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted allocations for instance e800f5e9-deb0-4592-aded-3768d238e311 [ 900.675322] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253094, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.795719] env[61936]: DEBUG nova.network.neutron [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7d859e6c-be30-4b94-9fb6-6af6858a2ba5", "address": "fa:16:3e:36:f9:55", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d859e6c-be", "ovs_interfaceid": "7d859e6c-be30-4b94-9fb6-6af6858a2ba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.820849] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565854} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.821147] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 10669072-a8e8-4262-b390-8bf31af6fafe/10669072-a8e8-4262-b390-8bf31af6fafe.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 900.821401] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.821678] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53799054-dc10-4e75-8669-d92046c78275 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.829448] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 900.829448] env[61936]: value = "task-1253095" [ 900.829448] env[61936]: _type = "Task" [ 900.829448] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.838089] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.082622] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Received event network-vif-plugged-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 901.082851] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Acquiring lock "2797bce9-f221-49b5-ab2b-42df3f347497-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.083076] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Lock "2797bce9-f221-49b5-ab2b-42df3f347497-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.083273] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Lock "2797bce9-f221-49b5-ab2b-42df3f347497-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.083427] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] No waiting events found dispatching network-vif-plugged-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 901.083594] env[61936]: WARNING nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Received unexpected event network-vif-plugged-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f for instance with vm_state building and task_state spawning. [ 901.083792] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Received event network-changed-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 901.083969] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Refreshing instance network info cache due to event network-changed-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 901.084177] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Acquiring lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.084327] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Acquired lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.084479] env[61936]: DEBUG nova.network.neutron [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Refreshing network info cache for port d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.088614] env[61936]: INFO nova.compute.manager [-] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Took 1.53 seconds to deallocate network for instance. [ 901.111695] env[61936]: INFO nova.compute.manager [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Took 31.24 seconds to build instance. [ 901.124430] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eba6bfda-7970-4d67-9c36-55f101a30dd6 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "e800f5e9-deb0-4592-aded-3768d238e311" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.945s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.177209] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253094, 'name': CreateVM_Task, 'duration_secs': 0.514708} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.177499] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.178713] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.178976] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.179442] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.179730] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adccb316-c8ba-4321-838a-c0e828d42128 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.185544] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 901.185544] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52218c2d-97bb-8e2a-bd00-935e33e29888" [ 901.185544] env[61936]: _type = "Task" [ 901.185544] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.194819] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52218c2d-97bb-8e2a-bd00-935e33e29888, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.298310] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.299022] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.299201] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.300131] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9400aa-dae4-4499-a473-3c9858814938 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.318207] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 901.318502] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.318703] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 901.318905] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.319066] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 901.319218] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 901.319435] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 901.319592] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 901.319757] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 901.319922] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 901.320111] env[61936]: DEBUG nova.virt.hardware [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 901.326634] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Reconfiguring VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 901.326989] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-480c0ce4-4bd4-420f-8b45-93e04cad9fc1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.349864] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.27807} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.351279] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.352210] env[61936]: DEBUG oslo_vmware.api [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 901.352210] env[61936]: value = "task-1253096" [ 901.352210] env[61936]: _type = "Task" [ 901.352210] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.352951] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8b7ba7-f0f4-4aaa-b949-0e15bf3654d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.364709] env[61936]: DEBUG oslo_vmware.api [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253096, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.382350] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 10669072-a8e8-4262-b390-8bf31af6fafe/10669072-a8e8-4262-b390-8bf31af6fafe.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.382683] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b7e5b4f-1850-4ee2-b45d-2c9d33cfd1dc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.406344] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 901.406344] env[61936]: value = "task-1253097" [ 901.406344] env[61936]: _type = "Task" [ 901.406344] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.415554] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.595807] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.615597] env[61936]: DEBUG oslo_concurrency.lockutils [None req-819e06ed-e491-4c8c-b984-ea81fbb7b8dc tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.612s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.701309] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52218c2d-97bb-8e2a-bd00-935e33e29888, 'name': SearchDatastore_Task, 'duration_secs': 0.051394} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.701685] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.702037] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.702350] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.702565] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.702908] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.703207] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e956660d-f821-4c91-8488-960c2a2fead7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.712905] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.713159] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.716041] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6d47f44-7296-47d7-a998-23018c0f149d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.722567] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 901.722567] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f5a83f-178d-da9f-54dd-65fdf9f9d099" [ 901.722567] env[61936]: _type = "Task" [ 901.722567] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.731439] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f5a83f-178d-da9f-54dd-65fdf9f9d099, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.869285] env[61936]: DEBUG oslo_vmware.api [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.917480] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253097, 'name': ReconfigVM_Task, 'duration_secs': 0.471911} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.917623] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 10669072-a8e8-4262-b390-8bf31af6fafe/10669072-a8e8-4262-b390-8bf31af6fafe.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.918442] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8b34040-dd28-4d6e-8502-5e580e1b08a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.927536] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 901.927536] env[61936]: value = "task-1253098" [ 901.927536] env[61936]: _type = "Task" [ 901.927536] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.942622] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253098, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.942622] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727eeb72-0b54-4afb-a2cd-ef637756a61d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.950066] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835cf618-ab68-4d60-a362-dcfd2dd23926 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.984337] env[61936]: DEBUG nova.network.neutron [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updated VIF entry in instance network info cache for port d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.984704] env[61936]: DEBUG nova.network.neutron [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updating instance_info_cache with network_info: [{"id": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "address": "fa:16:3e:bb:c3:ef", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0bdd1e2-85", "ovs_interfaceid": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.986352] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fba11f9-e4f9-4abc-9d4d-a25dfcfd13d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.995467] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba33f46-e24d-49d7-a3a8-b33e7d7c6b14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.016631] env[61936]: DEBUG nova.compute.provider_tree [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.233903] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f5a83f-178d-da9f-54dd-65fdf9f9d099, 'name': SearchDatastore_Task, 'duration_secs': 0.010773} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.234845] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4e98045-e57d-41c6-b0f9-93032ff77c91 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.241141] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 902.241141] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5255c131-d5cd-f79c-73f6-2c1eefbcf134" [ 902.241141] env[61936]: _type = "Task" [ 902.241141] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.250785] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5255c131-d5cd-f79c-73f6-2c1eefbcf134, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.369686] env[61936]: DEBUG oslo_vmware.api [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253096, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.374610] env[61936]: DEBUG nova.network.neutron [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Successfully updated port: db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.384658] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "628882ff-2f0f-49ae-8454-d3791fb55545" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.384904] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.438396] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253098, 'name': Rename_Task, 'duration_secs': 0.220752} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.438718] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.438976] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3142c256-d3f8-4b12-8415-eaa14cdce224 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.441956] env[61936]: DEBUG nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 902.442494] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 902.442709] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.442866] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 902.443137] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.443302] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 902.443474] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 902.443688] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 902.443850] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 902.444036] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 902.444208] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 902.444414] env[61936]: DEBUG nova.virt.hardware [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 902.445278] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8931e9c4-eb45-4aa1-924c-fb8949e5f8a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.449094] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 902.449094] env[61936]: value = "task-1253099" [ 902.449094] env[61936]: _type = "Task" [ 902.449094] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.456623] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88189408-6ab5-4dd6-b238-fcece188c95c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.464424] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.490434] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Releasing lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.490743] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received event network-vif-plugged-7d859e6c-be30-4b94-9fb6-6af6858a2ba5 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 902.490894] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.491177] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.491279] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.491593] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] No waiting events found dispatching network-vif-plugged-7d859e6c-be30-4b94-9fb6-6af6858a2ba5 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 902.491758] env[61936]: WARNING nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received unexpected event network-vif-plugged-7d859e6c-be30-4b94-9fb6-6af6858a2ba5 for instance with vm_state active and task_state None. [ 902.491934] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received event network-changed-7d859e6c-be30-4b94-9fb6-6af6858a2ba5 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 902.492254] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Refreshing instance network info cache due to event network-changed-7d859e6c-be30-4b94-9fb6-6af6858a2ba5. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 902.492429] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Acquiring lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.492881] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Acquired lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.492881] env[61936]: DEBUG nova.network.neutron [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Refreshing network info cache for port 7d859e6c-be30-4b94-9fb6-6af6858a2ba5 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.520851] env[61936]: DEBUG nova.scheduler.client.report [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 902.752487] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5255c131-d5cd-f79c-73f6-2c1eefbcf134, 'name': SearchDatastore_Task, 'duration_secs': 0.015288} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.752758] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.753026] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 2797bce9-f221-49b5-ab2b-42df3f347497/2797bce9-f221-49b5-ab2b-42df3f347497.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.753327] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3153d8b0-e36f-41ad-80e6-23a6de46ba6f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.760225] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 902.760225] env[61936]: value = "task-1253100" [ 902.760225] env[61936]: _type = "Task" [ 902.760225] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.768163] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.867567] env[61936]: DEBUG oslo_vmware.api [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253096, 'name': ReconfigVM_Task, 'duration_secs': 1.156359} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.868173] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.868465] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Reconfigured VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 902.879138] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquiring lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.879282] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquired lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.879462] env[61936]: DEBUG nova.network.neutron [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.887985] env[61936]: DEBUG nova.compute.utils [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.960403] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253099, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.028476] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.028476] env[61936]: DEBUG nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 903.030781] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.988s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.032105] env[61936]: DEBUG nova.objects.instance [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lazy-loading 'resources' on Instance uuid 77a59650-d28b-4862-bb35-6910e6f7d6c9 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.218595] env[61936]: DEBUG nova.network.neutron [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updated VIF entry in instance network info cache for port 7d859e6c-be30-4b94-9fb6-6af6858a2ba5. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.219074] env[61936]: DEBUG nova.network.neutron [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7d859e6c-be30-4b94-9fb6-6af6858a2ba5", "address": "fa:16:3e:36:f9:55", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d859e6c-be", "ovs_interfaceid": "7d859e6c-be30-4b94-9fb6-6af6858a2ba5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.273217] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253100, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.373327] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e8443860-6356-4f1f-9285-6ee3e67cc4a9 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-e15be950-1c32-4801-b4e4-a4d31ff0cfec-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.233s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.390362] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.421708] env[61936]: DEBUG nova.network.neutron [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.463860] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253099, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.534578] env[61936]: DEBUG nova.compute.utils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 903.536144] env[61936]: DEBUG nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 903.536273] env[61936]: DEBUG nova.network.neutron [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 903.605497] env[61936]: DEBUG nova.network.neutron [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updating instance_info_cache with network_info: [{"id": "db50ae88-a45b-406e-855a-7124a33ce9ee", "address": "fa:16:3e:18:eb:4d", "network": {"id": "899b1c80-9135-4ecb-bb5c-d5ef9f438c32", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1588461718-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2171cb70a814ecb986099f6149ffc89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb50ae88-a4", "ovs_interfaceid": "db50ae88-a45b-406e-855a-7124a33ce9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.610087] env[61936]: DEBUG nova.policy [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '288fd1d0d8e341999481fc62fe14195e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f1b64c32564b9cbb68e3799b456641', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.722425] env[61936]: DEBUG oslo_concurrency.lockutils [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] Releasing lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.722711] env[61936]: DEBUG nova.compute.manager [req-bcd9fc35-b669-4c6b-bc37-0924f1226f8f req-26786bed-c919-443f-a559-26581f143940 service nova] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Received event network-vif-deleted-9f132878-bd3a-45e4-a5a0-03051b7bfd0e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 903.775151] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.797789} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.775487] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 2797bce9-f221-49b5-ab2b-42df3f347497/2797bce9-f221-49b5-ab2b-42df3f347497.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.775769] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.776073] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1bbbb09b-abc5-47cb-9312-d5960b795e97 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.803083] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 903.803083] env[61936]: value = "task-1253101" [ 903.803083] env[61936]: _type = "Task" [ 903.803083] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.814966] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.854534] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10699716-ee9d-48cf-b4de-2137971a5f3b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.862502] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9917f659-3a58-48af-9308-574137aba54c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.894824] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95790875-a10d-4afd-9c90-089574c83196 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.904338] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53c9b22-b03b-4a46-bf18-5a43dcdd158a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.909048] env[61936]: DEBUG nova.network.neutron [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Successfully created port: 432e9691-4e6f-4550-a7f4-7396d05e96d7 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.924774] env[61936]: DEBUG nova.compute.provider_tree [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.961055] env[61936]: DEBUG oslo_vmware.api [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253099, 'name': PowerOnVM_Task, 'duration_secs': 1.398457} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.961549] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.961857] env[61936]: INFO nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Took 9.29 seconds to spawn the instance on the hypervisor. [ 903.962390] env[61936]: DEBUG nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 903.963365] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb9bb9f-8f6a-4f45-ad21-5c5d616352f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.046446] env[61936]: DEBUG nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 904.108396] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Releasing lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.108732] env[61936]: DEBUG nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance network_info: |[{"id": "db50ae88-a45b-406e-855a-7124a33ce9ee", "address": "fa:16:3e:18:eb:4d", "network": {"id": "899b1c80-9135-4ecb-bb5c-d5ef9f438c32", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1588461718-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2171cb70a814ecb986099f6149ffc89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb50ae88-a4", "ovs_interfaceid": "db50ae88-a45b-406e-855a-7124a33ce9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 904.109435] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:eb:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4aa1eda7-48b9-4fa2-af0b-94c718313af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'db50ae88-a45b-406e-855a-7124a33ce9ee', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.117208] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Creating folder: Project (c2171cb70a814ecb986099f6149ffc89). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 904.118397] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03481347-8a74-4c29-b689-8c035a67f3ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.134787] env[61936]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 904.134962] env[61936]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61936) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 904.135461] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Folder already exists: Project (c2171cb70a814ecb986099f6149ffc89). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 904.136048] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Creating folder: Instances. Parent ref: group-v269936. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 904.136412] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7804c7e7-31c9-449a-b5ab-6a2ad37c8b12 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.150328] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Created folder: Instances in parent group-v269936. [ 904.150328] env[61936]: DEBUG oslo.service.loopingcall [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.150750] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.150975] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba653c74-e4d8-4e85-8ef3-2523395abc0a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.171829] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.171829] env[61936]: value = "task-1253104" [ 904.171829] env[61936]: _type = "Task" [ 904.171829] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.184188] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253104, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.316873] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073169} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.317621] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.319023] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca251ac9-7588-4dfd-af52-9eceafd29986 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.346350] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 2797bce9-f221-49b5-ab2b-42df3f347497/2797bce9-f221-49b5-ab2b-42df3f347497.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.346691] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7059eff-5469-40c6-8503-941ea97bac71 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.369602] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 904.369602] env[61936]: value = "task-1253105" [ 904.369602] env[61936]: _type = "Task" [ 904.369602] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.378710] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.428810] env[61936]: DEBUG nova.scheduler.client.report [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 904.478569] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "628882ff-2f0f-49ae-8454-d3791fb55545" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.478847] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.479094] env[61936]: INFO nova.compute.manager [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Attaching volume 42cbbe71-2fab-4c61-9288-650b1dec5002 to /dev/sdb [ 904.493734] env[61936]: INFO nova.compute.manager [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Took 33.06 seconds to build instance. [ 904.525211] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192a2e70-80ae-460c-b13a-88182b1375ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.534573] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfb9f93-c466-4d67-898a-6934dc1c7726 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.558178] env[61936]: DEBUG nova.virt.block_device [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Updating existing volume attachment record: 344b91ef-2ebf-40ce-8eec-9ff1f453fbc7 {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 904.683043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.683043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.683227] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.683362] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.683516] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.685102] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253104, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.685635] env[61936]: INFO nova.compute.manager [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Terminating instance [ 904.700460] env[61936]: DEBUG nova.compute.manager [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Received event network-vif-plugged-db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 904.700657] env[61936]: DEBUG oslo_concurrency.lockutils [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] Acquiring lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.700862] env[61936]: DEBUG oslo_concurrency.lockutils [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.701140] env[61936]: DEBUG oslo_concurrency.lockutils [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.701625] env[61936]: DEBUG nova.compute.manager [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] No waiting events found dispatching network-vif-plugged-db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 904.701625] env[61936]: WARNING nova.compute.manager [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Received unexpected event network-vif-plugged-db50ae88-a45b-406e-855a-7124a33ce9ee for instance with vm_state building and task_state spawning. [ 904.701810] env[61936]: DEBUG nova.compute.manager [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Received event network-changed-db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 904.701810] env[61936]: DEBUG nova.compute.manager [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Refreshing instance network info cache due to event network-changed-db50ae88-a45b-406e-855a-7124a33ce9ee. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 904.701996] env[61936]: DEBUG oslo_concurrency.lockutils [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] Acquiring lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.702150] env[61936]: DEBUG oslo_concurrency.lockutils [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] Acquired lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.702308] env[61936]: DEBUG nova.network.neutron [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Refreshing network info cache for port db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.881218] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253105, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.937457] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.939942] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.794s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.941431] env[61936]: INFO nova.compute.claims [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.975059] env[61936]: INFO nova.scheduler.client.report [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleted allocations for instance 77a59650-d28b-4862-bb35-6910e6f7d6c9 [ 904.996559] env[61936]: DEBUG oslo_concurrency.lockutils [None req-90ec1ac2-73db-4060-8b2b-deed8e6d4977 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "10669072-a8e8-4262-b390-8bf31af6fafe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.237s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.062431] env[61936]: DEBUG nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 905.094235] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 905.095321] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.095746] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.096129] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.096554] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.097735] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 905.098079] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 905.098354] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 905.098629] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 905.098865] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 905.099110] env[61936]: DEBUG nova.virt.hardware [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 905.101168] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3456892b-2dae-4ff0-ae7c-83e5cf18a6d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.112185] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9639eb5-8bb7-4f07-be60-7cdbac530e6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.183384] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253104, 'name': CreateVM_Task, 'duration_secs': 0.630739} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.183608] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.184388] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269950', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'name': 'volume-612c642f-abe9-4886-ac2f-baadfe0dcb78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a12fd59-bacd-42b3-99c1-8910f349e2cb', 'attached_at': '', 'detached_at': '', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'serial': '612c642f-abe9-4886-ac2f-baadfe0dcb78'}, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'disk_bus': None, 'device_type': None, 'attachment_id': '65da5d25-fafb-4fe8-98d2-19c7a0097487', 'volume_type': None}], 'swap': None} {{(pid=61936) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 905.184621] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Root volume attach. Driver type: vmdk {{(pid=61936) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 905.185466] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb74ccf-f0d3-473b-8c99-d4cd020f5dd1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.188866] env[61936]: DEBUG nova.compute.manager [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 905.189073] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 905.189816] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5da0950-da32-4112-ab90-f60908f683d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.200015] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f08c90-35c8-4cac-9c82-dcff7761b4ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.202749] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.202984] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24c66bee-c3e1-4f5f-933e-d6c995619289 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.209694] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993825fe-ec92-478b-af5c-973dd73cec56 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.214126] env[61936]: DEBUG oslo_vmware.api [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 905.214126] env[61936]: value = "task-1253109" [ 905.214126] env[61936]: _type = "Task" [ 905.214126] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.219228] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-353ebce5-af41-4558-a34f-4fdc7ad33bac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.228870] env[61936]: DEBUG oslo_vmware.api [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253109, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.230527] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 905.230527] env[61936]: value = "task-1253110" [ 905.230527] env[61936]: _type = "Task" [ 905.230527] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.239389] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.383259] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253105, 'name': ReconfigVM_Task, 'duration_secs': 0.775748} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.383570] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 2797bce9-f221-49b5-ab2b-42df3f347497/2797bce9-f221-49b5-ab2b-42df3f347497.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.384626] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11031f43-1acb-4e89-9ed9-204a04f91453 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.393726] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 905.393726] env[61936]: value = "task-1253111" [ 905.393726] env[61936]: _type = "Task" [ 905.393726] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.407688] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253111, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.439160] env[61936]: DEBUG nova.compute.manager [req-86956209-ff61-4673-ab01-4d33cfd2e081 req-bc8c6171-e75c-40be-b53d-80e37f11507e service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Received event network-vif-plugged-432e9691-4e6f-4550-a7f4-7396d05e96d7 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 905.440089] env[61936]: DEBUG oslo_concurrency.lockutils [req-86956209-ff61-4673-ab01-4d33cfd2e081 req-bc8c6171-e75c-40be-b53d-80e37f11507e service nova] Acquiring lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.440089] env[61936]: DEBUG oslo_concurrency.lockutils [req-86956209-ff61-4673-ab01-4d33cfd2e081 req-bc8c6171-e75c-40be-b53d-80e37f11507e service nova] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.440089] env[61936]: DEBUG oslo_concurrency.lockutils [req-86956209-ff61-4673-ab01-4d33cfd2e081 req-bc8c6171-e75c-40be-b53d-80e37f11507e service nova] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.440089] env[61936]: DEBUG nova.compute.manager [req-86956209-ff61-4673-ab01-4d33cfd2e081 req-bc8c6171-e75c-40be-b53d-80e37f11507e service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] No waiting events found dispatching network-vif-plugged-432e9691-4e6f-4550-a7f4-7396d05e96d7 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 905.440288] env[61936]: WARNING nova.compute.manager [req-86956209-ff61-4673-ab01-4d33cfd2e081 req-bc8c6171-e75c-40be-b53d-80e37f11507e service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Received unexpected event network-vif-plugged-432e9691-4e6f-4550-a7f4-7396d05e96d7 for instance with vm_state building and task_state spawning. [ 905.468714] env[61936]: DEBUG nova.network.neutron [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Successfully updated port: 432e9691-4e6f-4550-a7f4-7396d05e96d7 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.489670] env[61936]: DEBUG oslo_concurrency.lockutils [None req-152edeb4-535b-409e-8e7f-69731fd85790 tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "77a59650-d28b-4862-bb35-6910e6f7d6c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.399s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.566765] env[61936]: DEBUG nova.compute.manager [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Received event network-changed-894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 905.567287] env[61936]: DEBUG nova.compute.manager [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Refreshing instance network info cache due to event network-changed-894b1eac-a089-4cf8-adbc-f2e646971faf. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 905.567430] env[61936]: DEBUG oslo_concurrency.lockutils [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] Acquiring lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.567615] env[61936]: DEBUG oslo_concurrency.lockutils [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] Acquired lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.567802] env[61936]: DEBUG nova.network.neutron [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Refreshing network info cache for port 894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.619412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-e15be950-1c32-4801-b4e4-a4d31ff0cfec-7d859e6c-be30-4b94-9fb6-6af6858a2ba5" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.619811] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-e15be950-1c32-4801-b4e4-a4d31ff0cfec-7d859e6c-be30-4b94-9fb6-6af6858a2ba5" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.630771] env[61936]: DEBUG nova.network.neutron [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updated VIF entry in instance network info cache for port db50ae88-a45b-406e-855a-7124a33ce9ee. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.631358] env[61936]: DEBUG nova.network.neutron [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updating instance_info_cache with network_info: [{"id": "db50ae88-a45b-406e-855a-7124a33ce9ee", "address": "fa:16:3e:18:eb:4d", "network": {"id": "899b1c80-9135-4ecb-bb5c-d5ef9f438c32", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1588461718-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2171cb70a814ecb986099f6149ffc89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb50ae88-a4", "ovs_interfaceid": "db50ae88-a45b-406e-855a-7124a33ce9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.727739] env[61936]: DEBUG oslo_vmware.api [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253109, 'name': PowerOffVM_Task, 'duration_secs': 0.174177} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.727739] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.728106] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 905.728106] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5dd14b59-a0b9-49be-a234-cf1fe50dddfc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.744719] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 42%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.859527] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 905.859802] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 905.860021] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleting the datastore file [datastore2] 9f6a1c42-e5a5-4f1d-9740-35dc6028183a {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.860336] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e73efae6-df48-41da-a2e1-5e57e5f2e8c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.869434] env[61936]: DEBUG oslo_vmware.api [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 905.869434] env[61936]: value = "task-1253113" [ 905.869434] env[61936]: _type = "Task" [ 905.869434] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.878745] env[61936]: DEBUG oslo_vmware.api [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.907905] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253111, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.971647] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.971934] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.972141] env[61936]: DEBUG nova.network.neutron [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.123995] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.123995] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.125154] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795a7012-e407-4907-be7e-2761aca6a296 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.145040] env[61936]: DEBUG oslo_concurrency.lockutils [req-1fe21acf-1d2d-4864-a451-9d54640521bb req-71a844c3-5d90-48ab-aaa7-c22fe3c5d1f5 service nova] Releasing lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.148874] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd40ca00-5326-4c52-a2d5-f619a310ef8d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.180844] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Reconfiguring VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 906.183979] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16d72f83-92af-4c3a-9b0e-7f2b04edaa87 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.207535] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 906.207535] env[61936]: value = "task-1253114" [ 906.207535] env[61936]: _type = "Task" [ 906.207535] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.220231] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.242398] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 54%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.345378] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b51d09-7039-4e1b-b56a-38453890bd7c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.355892] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-631c207a-8211-4244-862f-cf0323af3cc4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.395343] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff284b3a-db78-45d2-9cd9-8f92472f4c92 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.398694] env[61936]: DEBUG nova.network.neutron [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Updated VIF entry in instance network info cache for port 894b1eac-a089-4cf8-adbc-f2e646971faf. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.399274] env[61936]: DEBUG nova.network.neutron [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Updating instance_info_cache with network_info: [{"id": "894b1eac-a089-4cf8-adbc-f2e646971faf", "address": "fa:16:3e:f9:eb:61", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap894b1eac-a0", "ovs_interfaceid": "894b1eac-a089-4cf8-adbc-f2e646971faf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.408055] env[61936]: DEBUG oslo_vmware.api [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277578} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.408415] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.408639] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 906.408806] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 906.409013] env[61936]: INFO nova.compute.manager [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Took 1.22 seconds to destroy the instance on the hypervisor. [ 906.409325] env[61936]: DEBUG oslo.service.loopingcall [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.411765] env[61936]: DEBUG nova.compute.manager [-] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 906.411934] env[61936]: DEBUG nova.network.neutron [-] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 906.418920] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00290fd4-666c-491f-83a9-2ede9eac0073 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.423800] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253111, 'name': Rename_Task, 'duration_secs': 0.615874} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.425129] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.425784] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96147263-731e-4f08-b562-18b949dff636 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.437811] env[61936]: DEBUG nova.compute.provider_tree [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.440564] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 906.440564] env[61936]: value = "task-1253115" [ 906.440564] env[61936]: _type = "Task" [ 906.440564] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.454244] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.539515] env[61936]: DEBUG nova.network.neutron [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 906.724154] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.744471] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 67%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.756227] env[61936]: DEBUG nova.compute.manager [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Received event network-changed-432e9691-4e6f-4550-a7f4-7396d05e96d7 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 906.756227] env[61936]: DEBUG nova.compute.manager [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Refreshing instance network info cache due to event network-changed-432e9691-4e6f-4550-a7f4-7396d05e96d7. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 906.756227] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] Acquiring lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.902202] env[61936]: DEBUG oslo_concurrency.lockutils [req-7929c53f-b715-41f8-bfc6-319f01d68cf6 req-04db6bae-2c1f-4981-a0fa-8265fde42f7e service nova] Releasing lock "refresh_cache-10669072-a8e8-4262-b390-8bf31af6fafe" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.946918] env[61936]: DEBUG nova.scheduler.client.report [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 906.958259] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253115, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.964290] env[61936]: DEBUG nova.network.neutron [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance_info_cache with network_info: [{"id": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "address": "fa:16:3e:34:7d:9b", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap432e9691-4e", "ovs_interfaceid": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.221242] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.245573] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 82%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.454501] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.455048] env[61936]: DEBUG nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 907.457869] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253115, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.458449] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.044s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.458684] env[61936]: DEBUG nova.objects.instance [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lazy-loading 'resources' on Instance uuid fafc995e-4c93-4ca0-b078-24a1ae0ab427 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.466725] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.466986] env[61936]: DEBUG nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Instance network_info: |[{"id": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "address": "fa:16:3e:34:7d:9b", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap432e9691-4e", "ovs_interfaceid": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 907.467419] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] Acquired lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.467615] env[61936]: DEBUG nova.network.neutron [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Refreshing network info cache for port 432e9691-4e6f-4550-a7f4-7396d05e96d7 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.468761] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:7d:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b94712a6-b777-47dd-bc06-f9acfce2d936', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '432e9691-4e6f-4550-a7f4-7396d05e96d7', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.476235] env[61936]: DEBUG oslo.service.loopingcall [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.477554] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.477813] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f61df7f0-ac12-470d-a630-2aa554d35be2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.503212] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.503212] env[61936]: value = "task-1253117" [ 907.503212] env[61936]: _type = "Task" [ 907.503212] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.516425] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253117, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.559208] env[61936]: DEBUG nova.network.neutron [-] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.718756] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.742657] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 97%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.952583] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253115, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.960113] env[61936]: DEBUG nova.compute.utils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.961502] env[61936]: DEBUG nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 907.961619] env[61936]: DEBUG nova.network.neutron [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 908.014257] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253117, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.062336] env[61936]: INFO nova.compute.manager [-] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Took 1.65 seconds to deallocate network for instance. [ 908.076994] env[61936]: DEBUG nova.policy [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9bb41ed24f244345b83f2a16f4f10804', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f1b73bc1dea4df0a995655f3bd50d14', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.220662] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.245976] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 97%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.295568] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad553e0-ec36-42c0-a644-81dc890d2358 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.303795] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2565a8f0-176d-448a-9ec8-60b5f5a3b182 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.333734] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af6cf7d-ac3b-4078-bf39-57bbc096e6b6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.341315] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109d5a5e-41db-4412-b55e-616ca619678d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.354647] env[61936]: DEBUG nova.compute.provider_tree [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.453662] env[61936]: DEBUG oslo_vmware.api [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253115, 'name': PowerOnVM_Task, 'duration_secs': 1.957852} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.453941] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.454175] env[61936]: INFO nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Took 11.33 seconds to spawn the instance on the hypervisor. [ 908.454357] env[61936]: DEBUG nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.455195] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809f0942-2e08-4ee2-84e1-cdc3f14e71f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.467269] env[61936]: DEBUG nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 908.483118] env[61936]: DEBUG nova.network.neutron [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updated VIF entry in instance network info cache for port 432e9691-4e6f-4550-a7f4-7396d05e96d7. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.483521] env[61936]: DEBUG nova.network.neutron [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance_info_cache with network_info: [{"id": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "address": "fa:16:3e:34:7d:9b", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap432e9691-4e", "ovs_interfaceid": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.515651] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253117, 'name': CreateVM_Task, 'duration_secs': 0.68914} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.515891] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.516729] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.517252] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.517661] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.517966] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76041de8-c5af-426f-b431-0de59402d3b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.523983] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 908.523983] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52524ddb-230a-abe0-535b-c9a24eb2f2c1" [ 908.523983] env[61936]: _type = "Task" [ 908.523983] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.530555] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52524ddb-230a-abe0-535b-c9a24eb2f2c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.571098] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.660690] env[61936]: DEBUG nova.network.neutron [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Successfully created port: a6719762-3a94-4d4a-9fea-ff09306d59b3 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.719937] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.744940] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 97%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.786763] env[61936]: DEBUG nova.compute.manager [req-b3e4f12c-e1fa-4bdb-8fae-094239f04315 req-26a8d3ee-3a8b-49f9-b9b7-38c4f7f27ddb service nova] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Received event network-vif-deleted-691101d8-d631-42c1-80ce-2aba05214046 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 908.857536] env[61936]: DEBUG nova.scheduler.client.report [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 908.976918] env[61936]: INFO nova.compute.manager [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Took 35.06 seconds to build instance. [ 908.988405] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ba720b6-9c6f-4042-848d-43a8d4e3cf8b req-46326d22-27db-4306-a28c-da9e1cea6224 service nova] Releasing lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.034211] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52524ddb-230a-abe0-535b-c9a24eb2f2c1, 'name': SearchDatastore_Task, 'duration_secs': 0.011602} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.034613] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.034939] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.035234] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.035410] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.035633] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.036313] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dde56d57-a07d-4f78-83d5-fa73032006db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.057562] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.057749] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.058566] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-414dce56-a952-4d3f-8b9c-6c9b1f7ee6ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.063758] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 909.063758] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52221211-5c95-5a72-656b-4b566f89024d" [ 909.063758] env[61936]: _type = "Task" [ 909.063758] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.071084] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52221211-5c95-5a72-656b-4b566f89024d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.104446] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 909.104684] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269977', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'name': 'volume-42cbbe71-2fab-4c61-9288-650b1dec5002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '628882ff-2f0f-49ae-8454-d3791fb55545', 'attached_at': '', 'detached_at': '', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'serial': '42cbbe71-2fab-4c61-9288-650b1dec5002'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 909.105557] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81138555-39fc-4492-83c0-1f9af081639a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.120977] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11dea865-80e1-4e38-a74a-d130dee1cef9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.144796] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] volume-42cbbe71-2fab-4c61-9288-650b1dec5002/volume-42cbbe71-2fab-4c61-9288-650b1dec5002.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.145064] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f580140-5345-4228-8a37-2cf10b9ddaca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.162534] env[61936]: DEBUG oslo_vmware.api [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 909.162534] env[61936]: value = "task-1253118" [ 909.162534] env[61936]: _type = "Task" [ 909.162534] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.171732] env[61936]: DEBUG oslo_vmware.api [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253118, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.221272] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.247040] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task} progress is 98%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.364112] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.366650] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.718s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.366987] env[61936]: DEBUG nova.objects.instance [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lazy-loading 'resources' on Instance uuid c4f17d38-25a3-4630-8bd4-bb40fa6fddee {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.386675] env[61936]: INFO nova.scheduler.client.report [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Deleted allocations for instance fafc995e-4c93-4ca0-b078-24a1ae0ab427 [ 909.475654] env[61936]: DEBUG nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 909.478242] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6d563319-abac-40fd-9d53-f907dab878fb tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.657s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.502868] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 909.503034] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.503098] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 909.503284] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.503490] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 909.503651] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 909.503797] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 909.503913] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 909.504103] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 909.504246] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 909.504933] env[61936]: DEBUG nova.virt.hardware [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 909.505386] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d82e7f8-0f90-46c3-8f9e-34a8322470fa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.514412] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f151618-3766-4fd6-bf8d-9d179b8a819f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.575096] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52221211-5c95-5a72-656b-4b566f89024d, 'name': SearchDatastore_Task, 'duration_secs': 0.041301} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.575950] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffc61925-47ce-4ac3-a6c6-33b87e72c58c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.581582] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 909.581582] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520467d2-3c5e-1f82-35f9-102ba42ba00f" [ 909.581582] env[61936]: _type = "Task" [ 909.581582] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.589413] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520467d2-3c5e-1f82-35f9-102ba42ba00f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.671947] env[61936]: DEBUG oslo_vmware.api [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253118, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.722559] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.745571] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253110, 'name': RelocateVM_Task, 'duration_secs': 4.338538} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.745919] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 909.746022] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269950', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'name': 'volume-612c642f-abe9-4886-ac2f-baadfe0dcb78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a12fd59-bacd-42b3-99c1-8910f349e2cb', 'attached_at': '', 'detached_at': '', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'serial': '612c642f-abe9-4886-ac2f-baadfe0dcb78'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 909.746888] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25e52c8-8ecf-4e75-bd26-7061138bbde2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.762939] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdea4e6-fe96-4de4-9dcc-5c20a3b6668c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.785911] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] volume-612c642f-abe9-4886-ac2f-baadfe0dcb78/volume-612c642f-abe9-4886-ac2f-baadfe0dcb78.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.786446] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54e790ce-196c-47ef-a123-0bc62384e137 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.806984] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 909.806984] env[61936]: value = "task-1253119" [ 909.806984] env[61936]: _type = "Task" [ 909.806984] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.815555] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.898021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1a0d862e-8ac5-4bcd-84a3-8510d58fe2ac tempest-ListServersNegativeTestJSON-1474759816 tempest-ListServersNegativeTestJSON-1474759816-project-member] Lock "fafc995e-4c93-4ca0-b078-24a1ae0ab427" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.595s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.094604] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520467d2-3c5e-1f82-35f9-102ba42ba00f, 'name': SearchDatastore_Task, 'duration_secs': 0.027268} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.094604] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.094855] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 9bd97f84-7c21-44ad-9ed0-d7e4097e264e/9bd97f84-7c21-44ad-9ed0-d7e4097e264e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.095143] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05be62f6-e080-4151-a69e-82c15768bdd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.101496] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 910.101496] env[61936]: value = "task-1253120" [ 910.101496] env[61936]: _type = "Task" [ 910.101496] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.112626] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253120, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.171850] env[61936]: DEBUG oslo_vmware.api [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253118, 'name': ReconfigVM_Task, 'duration_secs': 0.812728} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.172172] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Reconfigured VM instance instance-00000049 to attach disk [datastore1] volume-42cbbe71-2fab-4c61-9288-650b1dec5002/volume-42cbbe71-2fab-4c61-9288-650b1dec5002.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.181186] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94840751-c18d-42c5-aeaf-27bfd7bf5ae3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.199791] env[61936]: DEBUG oslo_vmware.api [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 910.199791] env[61936]: value = "task-1253121" [ 910.199791] env[61936]: _type = "Task" [ 910.199791] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.208857] env[61936]: DEBUG oslo_vmware.api [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253121, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.221716] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.225194] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b70c21b-c8f4-4fc0-a451-f7b0803eadc1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.232132] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3fb48d-d945-4e59-ba43-a10117776b19 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.268939] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c137fd0-c31d-486a-a582-1f4e6ba1aa00 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.276865] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60647dc8-0ec1-4069-bfe7-4304d7b3e917 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.292267] env[61936]: DEBUG nova.compute.provider_tree [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.317811] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253119, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.618599] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253120, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.713377] env[61936]: DEBUG oslo_vmware.api [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253121, 'name': ReconfigVM_Task, 'duration_secs': 0.150801} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.713722] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269977', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'name': 'volume-42cbbe71-2fab-4c61-9288-650b1dec5002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '628882ff-2f0f-49ae-8454-d3791fb55545', 'attached_at': '', 'detached_at': '', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'serial': '42cbbe71-2fab-4c61-9288-650b1dec5002'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 910.733486] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.733486] env[61936]: DEBUG nova.network.neutron [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Successfully updated port: a6719762-3a94-4d4a-9fea-ff09306d59b3 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.797174] env[61936]: DEBUG nova.scheduler.client.report [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 910.818339] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253119, 'name': ReconfigVM_Task, 'duration_secs': 0.525667} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.819306] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Reconfigured VM instance instance-0000004d to attach disk [datastore1] volume-612c642f-abe9-4886-ac2f-baadfe0dcb78/volume-612c642f-abe9-4886-ac2f-baadfe0dcb78.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.824809] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8baef489-abd0-45df-8388-3451621cb8a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.841627] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 910.841627] env[61936]: value = "task-1253122" [ 910.841627] env[61936]: _type = "Task" [ 910.841627] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.851941] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253122, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.908016] env[61936]: DEBUG nova.compute.manager [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Received event network-changed-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 910.908016] env[61936]: DEBUG nova.compute.manager [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Refreshing instance network info cache due to event network-changed-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 910.908016] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Acquiring lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.908460] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Acquired lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.908883] env[61936]: DEBUG nova.network.neutron [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Refreshing network info cache for port d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 911.002029] env[61936]: DEBUG oslo_concurrency.lockutils [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "2797bce9-f221-49b5-ab2b-42df3f347497" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.002588] env[61936]: DEBUG oslo_concurrency.lockutils [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.002588] env[61936]: INFO nova.compute.manager [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Rebooting instance [ 911.116724] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253120, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611821} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.116979] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 9bd97f84-7c21-44ad-9ed0-d7e4097e264e/9bd97f84-7c21-44ad-9ed0-d7e4097e264e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.120492] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.120792] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73e30c96-5632-4158-8d4d-883aa452cc3b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.127306] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 911.127306] env[61936]: value = "task-1253123" [ 911.127306] env[61936]: _type = "Task" [ 911.127306] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.140514] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253123, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.233235] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.236306] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.236360] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.236596] env[61936]: DEBUG nova.network.neutron [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.300922] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.934s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.303686] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.842s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.304203] env[61936]: DEBUG nova.objects.instance [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 911.327492] env[61936]: INFO nova.scheduler.client.report [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Deleted allocations for instance c4f17d38-25a3-4630-8bd4-bb40fa6fddee [ 911.352864] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253122, 'name': ReconfigVM_Task, 'duration_secs': 0.336812} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.353186] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269950', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'name': 'volume-612c642f-abe9-4886-ac2f-baadfe0dcb78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a12fd59-bacd-42b3-99c1-8910f349e2cb', 'attached_at': '', 'detached_at': '', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'serial': '612c642f-abe9-4886-ac2f-baadfe0dcb78'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 911.353729] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4823cc67-0f6f-4d58-bafe-88b90d151007 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.361071] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 911.361071] env[61936]: value = "task-1253124" [ 911.361071] env[61936]: _type = "Task" [ 911.361071] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.369276] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253124, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.534497] env[61936]: DEBUG oslo_concurrency.lockutils [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.637667] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253123, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065738} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.638024] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.638813] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ccc35e-7668-42a0-8af5-9f7ddf18de17 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.663137] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 9bd97f84-7c21-44ad-9ed0-d7e4097e264e/9bd97f84-7c21-44ad-9ed0-d7e4097e264e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.663486] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-441895cc-a130-4784-a6f5-339e6b7608f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.685898] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 911.685898] env[61936]: value = "task-1253125" [ 911.685898] env[61936]: _type = "Task" [ 911.685898] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.694110] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253125, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.728278] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.774356] env[61936]: DEBUG nova.network.neutron [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updated VIF entry in instance network info cache for port d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 911.774774] env[61936]: DEBUG nova.network.neutron [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updating instance_info_cache with network_info: [{"id": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "address": "fa:16:3e:bb:c3:ef", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0bdd1e2-85", "ovs_interfaceid": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.780062] env[61936]: DEBUG nova.objects.instance [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'flavor' on Instance uuid 628882ff-2f0f-49ae-8454-d3791fb55545 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.792925] env[61936]: DEBUG nova.network.neutron [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 911.835922] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ba2d67a-bc9d-468a-aa02-271945beb62e tempest-ServersTestMultiNic-493909654 tempest-ServersTestMultiNic-493909654-project-member] Lock "c4f17d38-25a3-4630-8bd4-bb40fa6fddee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.847s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.871744] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253124, 'name': Rename_Task, 'duration_secs': 0.210465} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.872021] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.872280] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07d07b69-a325-48b1-b26d-55ec29cb3d2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.879578] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 911.879578] env[61936]: value = "task-1253126" [ 911.879578] env[61936]: _type = "Task" [ 911.879578] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.888254] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253126, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.074601] env[61936]: DEBUG nova.network.neutron [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Updating instance_info_cache with network_info: [{"id": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "address": "fa:16:3e:0e:6b:01", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6719762-3a", "ovs_interfaceid": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.201512] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253125, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.228230] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.278342] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Releasing lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.278663] env[61936]: DEBUG nova.compute.manager [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Received event network-vif-plugged-a6719762-3a94-4d4a-9fea-ff09306d59b3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 912.278879] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Acquiring lock "fa845f8e-957c-4c0b-a647-190c32989dcd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.279091] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.279594] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.279594] env[61936]: DEBUG nova.compute.manager [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] No waiting events found dispatching network-vif-plugged-a6719762-3a94-4d4a-9fea-ff09306d59b3 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 912.279718] env[61936]: WARNING nova.compute.manager [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Received unexpected event network-vif-plugged-a6719762-3a94-4d4a-9fea-ff09306d59b3 for instance with vm_state building and task_state spawning. [ 912.279785] env[61936]: DEBUG nova.compute.manager [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Received event network-changed-a6719762-3a94-4d4a-9fea-ff09306d59b3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 912.279946] env[61936]: DEBUG nova.compute.manager [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Refreshing instance network info cache due to event network-changed-a6719762-3a94-4d4a-9fea-ff09306d59b3. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 912.280221] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Acquiring lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.280348] env[61936]: DEBUG oslo_concurrency.lockutils [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquired lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.280511] env[61936]: DEBUG nova.network.neutron [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.283699] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f217aff1-d761-4539-883d-e83a6660ac94 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.805s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.314375] env[61936]: DEBUG oslo_concurrency.lockutils [None req-352fb8aa-543c-4d63-ab0b-0049f67df71b tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.315455] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.446s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.318414] env[61936]: DEBUG nova.objects.instance [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lazy-loading 'resources' on Instance uuid efdce04f-b0aa-49a6-94b8-1c960ac33fea {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.389924] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253126, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.578340] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.579484] env[61936]: DEBUG nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Instance network_info: |[{"id": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "address": "fa:16:3e:0e:6b:01", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6719762-3a", "ovs_interfaceid": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 912.579887] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Acquired lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.580168] env[61936]: DEBUG nova.network.neutron [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Refreshing network info cache for port a6719762-3a94-4d4a-9fea-ff09306d59b3 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 912.581716] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:6b:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6719762-3a94-4d4a-9fea-ff09306d59b3', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.593193] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating folder: Project (6f1b73bc1dea4df0a995655f3bd50d14). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.594630] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-adc93371-6b24-429c-a601-4123ff8677aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.606319] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created folder: Project (6f1b73bc1dea4df0a995655f3bd50d14) in parent group-v269874. [ 912.606547] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating folder: Instances. Parent ref: group-v269979. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 912.606806] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ae275a0-7fc4-4325-a4e2-2398b260e883 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.616087] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created folder: Instances in parent group-v269979. [ 912.616473] env[61936]: DEBUG oslo.service.loopingcall [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.616758] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.617080] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62d97323-d979-45ee-bf81-bf20e39a3f25 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.637611] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.637611] env[61936]: value = "task-1253129" [ 912.637611] env[61936]: _type = "Task" [ 912.637611] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.645092] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253129, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.696096] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253125, 'name': ReconfigVM_Task, 'duration_secs': 0.625455} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.696381] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 9bd97f84-7c21-44ad-9ed0-d7e4097e264e/9bd97f84-7c21-44ad-9ed0-d7e4097e264e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.697102] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da7987e8-95f7-4539-9ded-e9b6d17bc820 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.702854] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 912.702854] env[61936]: value = "task-1253130" [ 912.702854] env[61936]: _type = "Task" [ 912.702854] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.715761] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253130, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.727920] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.901034] env[61936]: DEBUG oslo_vmware.api [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253126, 'name': PowerOnVM_Task, 'duration_secs': 0.626505} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.901034] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.901034] env[61936]: INFO nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Took 10.46 seconds to spawn the instance on the hypervisor. [ 912.901333] env[61936]: DEBUG nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 912.902148] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3014f174-217d-4e59-b8f9-48ff069fbc98 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.923288] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "628882ff-2f0f-49ae-8454-d3791fb55545" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.923544] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.923757] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "628882ff-2f0f-49ae-8454-d3791fb55545-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.923940] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.924243] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.927463] env[61936]: INFO nova.compute.manager [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Terminating instance [ 913.090022] env[61936]: DEBUG nova.network.neutron [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updating instance_info_cache with network_info: [{"id": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "address": "fa:16:3e:bb:c3:ef", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0bdd1e2-85", "ovs_interfaceid": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.150864] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253129, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.152668] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3994ac5c-dbb7-496f-892c-b07f2fe1a75d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.159587] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fca5719-24b4-4cf2-8c7b-17d74cfd422d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.194237] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087d2ba5-09ed-4b9c-87ec-d112a4a3654d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.202623] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2db8db-6fce-451f-9517-39867ce86ebf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.219052] env[61936]: DEBUG nova.compute.provider_tree [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.223582] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253130, 'name': Rename_Task, 'duration_secs': 0.275071} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.228044] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.228390] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a9f7a7e-d4ac-4140-a8ce-7d3667a5b182 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.236202] env[61936]: DEBUG oslo_vmware.api [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253114, 'name': ReconfigVM_Task, 'duration_secs': 6.540519} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.238017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.238293] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Reconfigured VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 913.240366] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 913.240366] env[61936]: value = "task-1253131" [ 913.240366] env[61936]: _type = "Task" [ 913.240366] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.249347] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.424080] env[61936]: INFO nova.compute.manager [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Took 36.20 seconds to build instance. [ 913.436519] env[61936]: DEBUG nova.compute.manager [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 913.436780] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.437152] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1889ee66-f8a0-4d44-b9f3-6c2dc5fc3b53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.444765] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 913.444765] env[61936]: value = "task-1253132" [ 913.444765] env[61936]: _type = "Task" [ 913.444765] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.456464] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.593786] env[61936]: DEBUG oslo_concurrency.lockutils [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Releasing lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.652122] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253129, 'name': CreateVM_Task, 'duration_secs': 0.526246} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.652122] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.652747] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.652921] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.653272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.653537] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8af8ed71-9b24-4022-b0dd-355c402b4daf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.658429] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 913.658429] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52360af0-c0c1-e8a7-569c-defec9365cc7" [ 913.658429] env[61936]: _type = "Task" [ 913.658429] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.666833] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52360af0-c0c1-e8a7-569c-defec9365cc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.726128] env[61936]: DEBUG nova.scheduler.client.report [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 913.753324] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253131, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.839816] env[61936]: DEBUG nova.network.neutron [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Updated VIF entry in instance network info cache for port a6719762-3a94-4d4a-9fea-ff09306d59b3. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 913.839816] env[61936]: DEBUG nova.network.neutron [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Updating instance_info_cache with network_info: [{"id": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "address": "fa:16:3e:0e:6b:01", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6719762-3a", "ovs_interfaceid": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.924266] env[61936]: DEBUG oslo_concurrency.lockutils [None req-06b843b6-9897-43b6-908e-9f1d22c4bf4e tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.928s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.964023] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253132, 'name': PowerOffVM_Task, 'duration_secs': 0.17666} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.964023] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.964023] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 913.964023] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269977', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'name': 'volume-42cbbe71-2fab-4c61-9288-650b1dec5002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '628882ff-2f0f-49ae-8454-d3791fb55545', 'attached_at': '', 'detached_at': '', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'serial': '42cbbe71-2fab-4c61-9288-650b1dec5002'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 913.964023] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089f835b-4ecb-4483-bf09-f8f1b6a585ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.989479] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572efbff-4401-47cc-b609-750b2ca59944 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.999812] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6400d368-774b-45a8-a57b-7f62238d596b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.010973] env[61936]: DEBUG nova.compute.manager [req-9a23ede7-8d3c-4bd0-bf9a-1a5e6121829e req-4396667f-68aa-4564-8d21-fb0e442160a1 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received event network-vif-deleted-7d859e6c-be30-4b94-9fb6-6af6858a2ba5 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 914.010973] env[61936]: INFO nova.compute.manager [req-9a23ede7-8d3c-4bd0-bf9a-1a5e6121829e req-4396667f-68aa-4564-8d21-fb0e442160a1 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Neutron deleted interface 7d859e6c-be30-4b94-9fb6-6af6858a2ba5; detaching it from the instance and deleting it from the info cache [ 914.010973] env[61936]: DEBUG nova.network.neutron [req-9a23ede7-8d3c-4bd0-bf9a-1a5e6121829e req-4396667f-68aa-4564-8d21-fb0e442160a1 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.031629] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fee8478-8980-41e5-97fd-395e046fba3e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.051698] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] The volume has not been displaced from its original location: [datastore1] volume-42cbbe71-2fab-4c61-9288-650b1dec5002/volume-42cbbe71-2fab-4c61-9288-650b1dec5002.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 914.057450] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 914.057881] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1bc8a87-5ee6-4c96-a6c5-5d6f3caed256 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.078548] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 914.078548] env[61936]: value = "task-1253133" [ 914.078548] env[61936]: _type = "Task" [ 914.078548] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.086544] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253133, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.098518] env[61936]: DEBUG nova.compute.manager [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.099679] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e712d39f-cf59-4ecd-ad3f-a9d3cb4034eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.168781] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52360af0-c0c1-e8a7-569c-defec9365cc7, 'name': SearchDatastore_Task, 'duration_secs': 0.009831} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.169099] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.169341] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.169576] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.169726] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.169917] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.170560] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f154f4ce-f123-46cb-a4fa-a87ca5bf9165 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.179799] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.179916] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.180667] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b49bf1e2-2a2a-43b8-b666-e4314112ccb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.186308] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 914.186308] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b6971-61e3-0ba4-decf-82b314b6d192" [ 914.186308] env[61936]: _type = "Task" [ 914.186308] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.194242] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b6971-61e3-0ba4-decf-82b314b6d192, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.232899] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.916s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.235213] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.258s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.235487] env[61936]: DEBUG nova.objects.instance [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lazy-loading 'resources' on Instance uuid 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.254183] env[61936]: DEBUG oslo_vmware.api [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253131, 'name': PowerOnVM_Task, 'duration_secs': 0.638136} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.254831] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.254831] env[61936]: INFO nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Took 9.19 seconds to spawn the instance on the hypervisor. [ 914.254979] env[61936]: DEBUG nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.257729] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fb593d-7bd9-4398-a72b-bab8cc23d722 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.268707] env[61936]: INFO nova.scheduler.client.report [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted allocations for instance efdce04f-b0aa-49a6-94b8-1c960ac33fea [ 914.342072] env[61936]: DEBUG oslo_concurrency.lockutils [req-50390a67-b9b8-4af9-babb-a5735ae9db3e req-6d35ac4b-ea06-4d56-9172-7736d7efcc6f service nova] Releasing lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.514920] env[61936]: DEBUG oslo_concurrency.lockutils [req-9a23ede7-8d3c-4bd0-bf9a-1a5e6121829e req-4396667f-68aa-4564-8d21-fb0e442160a1 service nova] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.514920] env[61936]: DEBUG oslo_concurrency.lockutils [req-9a23ede7-8d3c-4bd0-bf9a-1a5e6121829e req-4396667f-68aa-4564-8d21-fb0e442160a1 service nova] Acquired lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.514920] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cb3e3c-88bf-484f-bba8-d7a9a179c2f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.535086] env[61936]: DEBUG oslo_concurrency.lockutils [req-9a23ede7-8d3c-4bd0-bf9a-1a5e6121829e req-4396667f-68aa-4564-8d21-fb0e442160a1 service nova] Releasing lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.535459] env[61936]: WARNING nova.compute.manager [req-9a23ede7-8d3c-4bd0-bf9a-1a5e6121829e req-4396667f-68aa-4564-8d21-fb0e442160a1 service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Detach interface failed, port_id=7d859e6c-be30-4b94-9fb6-6af6858a2ba5, reason: No device with interface-id 7d859e6c-be30-4b94-9fb6-6af6858a2ba5 exists on VM: nova.exception.NotFound: No device with interface-id 7d859e6c-be30-4b94-9fb6-6af6858a2ba5 exists on VM [ 914.596605] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253133, 'name': ReconfigVM_Task, 'duration_secs': 0.493851} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.596605] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 914.602171] env[61936]: DEBUG nova.compute.manager [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Received event network-changed-db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 914.602401] env[61936]: DEBUG nova.compute.manager [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Refreshing instance network info cache due to event network-changed-db50ae88-a45b-406e-855a-7124a33ce9ee. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 914.602804] env[61936]: DEBUG oslo_concurrency.lockutils [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] Acquiring lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.602804] env[61936]: DEBUG oslo_concurrency.lockutils [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] Acquired lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.603269] env[61936]: DEBUG nova.network.neutron [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Refreshing network info cache for port db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 914.605704] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebbd5cd5-da35-4337-8984-06ff357cc375 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.634305] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 914.634305] env[61936]: value = "task-1253134" [ 914.634305] env[61936]: _type = "Task" [ 914.634305] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.647478] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253134, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.699432] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b6971-61e3-0ba4-decf-82b314b6d192, 'name': SearchDatastore_Task, 'duration_secs': 0.009497} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.700633] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f4389e6-53e3-4b4b-9728-bc8983ebe74c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.706366] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 914.706366] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522d01bf-cbf5-a834-3306-3f69d8d42bc8" [ 914.706366] env[61936]: _type = "Task" [ 914.706366] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.715747] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522d01bf-cbf5-a834-3306-3f69d8d42bc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.796182] env[61936]: DEBUG oslo_concurrency.lockutils [None req-952a48d2-cd48-4fb2-b021-3fadbe16e75e tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "efdce04f-b0aa-49a6-94b8-1c960ac33fea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.840s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.798274] env[61936]: INFO nova.compute.manager [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Took 34.82 seconds to build instance. [ 914.959180] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.959454] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.959644] env[61936]: DEBUG nova.network.neutron [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.099606] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ca115e-bbdf-4533-8acd-26432f6db1fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.110116] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ce6dd7-98a0-44c1-b84f-a6e06d4fd44f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.937658] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.937658] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.937934] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.937934] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.938102] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.939801] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9ed4fb77-d283-4ab5-b988-8ee11b39c290 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.048s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.942976] env[61936]: INFO nova.compute.manager [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Terminating instance [ 915.947786] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f84e87-5a25-45cd-a986-b803435d2114 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.955758] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3bf352-cead-401b-9ca9-9d89943eceac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.973794] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Doing hard reboot of VM {{(pid=61936) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 915.974165] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522d01bf-cbf5-a834-3306-3f69d8d42bc8, 'name': SearchDatastore_Task, 'duration_secs': 0.010229} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.974529] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253134, 'name': ReconfigVM_Task, 'duration_secs': 0.268721} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.974604] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-d5dd0b7a-3052-47d1-82dd-95f74ec63418 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.977184] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd294155-c31f-4d35-a08c-730ce62f1e7c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.981434] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.981701] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/fa845f8e-957c-4c0b-a647-190c32989dcd.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.982026] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269977', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'name': 'volume-42cbbe71-2fab-4c61-9288-650b1dec5002', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '628882ff-2f0f-49ae-8454-d3791fb55545', 'attached_at': '', 'detached_at': '', 'volume_id': '42cbbe71-2fab-4c61-9288-650b1dec5002', 'serial': '42cbbe71-2fab-4c61-9288-650b1dec5002'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 915.982349] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 915.982624] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e31a9d6-34c4-4d9b-bcb7-6ac2839110ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.984967] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29550f0d-0fbd-459d-8d3d-4cd83cec9274 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.998541] env[61936]: DEBUG nova.compute.provider_tree [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.004199] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 916.004199] env[61936]: value = "task-1253136" [ 916.004199] env[61936]: _type = "Task" [ 916.004199] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.004598] env[61936]: DEBUG oslo_vmware.api [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 916.004598] env[61936]: value = "task-1253135" [ 916.004598] env[61936]: _type = "Task" [ 916.004598] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.004913] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.005630] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02a272ed-52db-49f4-80c9-58333130e8aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.023149] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253136, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.026265] env[61936]: DEBUG oslo_vmware.api [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253135, 'name': ResetVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.079482] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.079482] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.082021] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleting the datastore file [datastore2] 628882ff-2f0f-49ae-8454-d3791fb55545 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.082021] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0dd4a0bb-ddea-444e-bc56-7ec8ff000f06 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.087501] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 916.087501] env[61936]: value = "task-1253138" [ 916.087501] env[61936]: _type = "Task" [ 916.087501] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.094526] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253138, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.248443] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "1d25e242-3542-4707-9112-3711e17df577" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.248843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "1d25e242-3542-4707-9112-3711e17df577" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.308204] env[61936]: DEBUG nova.network.neutron [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updated VIF entry in instance network info cache for port db50ae88-a45b-406e-855a-7124a33ce9ee. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.308204] env[61936]: DEBUG nova.network.neutron [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updating instance_info_cache with network_info: [{"id": "db50ae88-a45b-406e-855a-7124a33ce9ee", "address": "fa:16:3e:18:eb:4d", "network": {"id": "899b1c80-9135-4ecb-bb5c-d5ef9f438c32", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1588461718-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.143", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2171cb70a814ecb986099f6149ffc89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb50ae88-a4", "ovs_interfaceid": "db50ae88-a45b-406e-855a-7124a33ce9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.439756] env[61936]: DEBUG nova.network.neutron [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [{"id": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "address": "fa:16:3e:0e:ca:c5", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12397fd2-b0", "ovs_interfaceid": "12397fd2-b04a-4019-b4fb-ca0f7b7fc152", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.464780] env[61936]: DEBUG nova.compute.manager [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 916.464780] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.464780] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e16acaa-5cf0-4563-8bda-02a0611284dc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.477135] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.477135] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01413e53-43d3-4f49-aecb-1d6bf370c16b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.482894] env[61936]: DEBUG oslo_vmware.api [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 916.482894] env[61936]: value = "task-1253139" [ 916.482894] env[61936]: _type = "Task" [ 916.482894] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.493251] env[61936]: DEBUG oslo_vmware.api [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.506722] env[61936]: DEBUG nova.scheduler.client.report [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 916.523218] env[61936]: DEBUG oslo_vmware.api [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253135, 'name': ResetVM_Task, 'duration_secs': 0.115831} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.526088] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Did hard reboot of VM {{(pid=61936) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 916.527149] env[61936]: DEBUG nova.compute.manager [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 916.527149] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253136, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.527458] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7f61ab-87dd-43ad-826a-7d6a19581f3e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.597075] env[61936]: DEBUG oslo_vmware.api [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253138, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.379743} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.602531] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.602531] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 916.602531] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 916.602531] env[61936]: INFO nova.compute.manager [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Took 3.16 seconds to destroy the instance on the hypervisor. [ 916.602531] env[61936]: DEBUG oslo.service.loopingcall [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.602531] env[61936]: DEBUG nova.compute.manager [-] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 916.602531] env[61936]: DEBUG nova.network.neutron [-] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 916.754869] env[61936]: DEBUG nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 916.810988] env[61936]: DEBUG oslo_concurrency.lockutils [req-92497e71-8486-4a1e-a755-8497e2285ed3 req-2c703753-86e2-45e5-95f2-2da1c05202d8 service nova] Releasing lock "refresh_cache-1a12fd59-bacd-42b3-99c1-8910f349e2cb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.941765] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-e15be950-1c32-4801-b4e4-a4d31ff0cfec" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.954735] env[61936]: DEBUG nova.compute.manager [req-fe112b98-05e5-40a3-98f6-eddf5d41db85 req-1bbde783-7259-47c3-8105-b258f984fd56 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Received event network-vif-deleted-39c4ceec-fef3-4d7d-8a6b-27fc061802a0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 916.955018] env[61936]: INFO nova.compute.manager [req-fe112b98-05e5-40a3-98f6-eddf5d41db85 req-1bbde783-7259-47c3-8105-b258f984fd56 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Neutron deleted interface 39c4ceec-fef3-4d7d-8a6b-27fc061802a0; detaching it from the instance and deleting it from the info cache [ 916.955562] env[61936]: DEBUG nova.network.neutron [req-fe112b98-05e5-40a3-98f6-eddf5d41db85 req-1bbde783-7259-47c3-8105-b258f984fd56 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.993644] env[61936]: DEBUG oslo_vmware.api [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253139, 'name': PowerOffVM_Task, 'duration_secs': 0.210865} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.993922] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.994102] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.994360] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d574819-a95c-43f7-9e18-951994ab934c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.020682] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.786s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.022894] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253136, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532884} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.024354] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.493s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.024354] env[61936]: DEBUG nova.objects.instance [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lazy-loading 'resources' on Instance uuid aa37292e-8499-46aa-8021-cb8d88b5a35a {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.024822] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/fa845f8e-957c-4c0b-a647-190c32989dcd.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.025063] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.025323] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a969a860-6935-4bec-82d0-c132ff0d09ee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.033237] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 917.033237] env[61936]: value = "task-1253141" [ 917.033237] env[61936]: _type = "Task" [ 917.033237] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.043769] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253141, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.044353] env[61936]: DEBUG oslo_concurrency.lockutils [None req-132b673f-9782-429d-88b4-978bd057c099 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 6.042s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.054015] env[61936]: INFO nova.scheduler.client.report [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted allocations for instance 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e [ 917.093025] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.093025] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.093025] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleting the datastore file [datastore1] e15be950-1c32-4801-b4e4-a4d31ff0cfec {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.093025] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42d6948f-3e07-4e91-a51b-0c4785b62682 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.101019] env[61936]: DEBUG oslo_vmware.api [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 917.101019] env[61936]: value = "task-1253142" [ 917.101019] env[61936]: _type = "Task" [ 917.101019] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.110166] env[61936]: DEBUG oslo_vmware.api [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253142, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.281116] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.429580] env[61936]: DEBUG nova.network.neutron [-] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.447121] env[61936]: DEBUG oslo_concurrency.lockutils [None req-230b276a-28a4-4761-b232-f8d6fed560d6 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-e15be950-1c32-4801-b4e4-a4d31ff0cfec-7d859e6c-be30-4b94-9fb6-6af6858a2ba5" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.827s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.459508] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8dba7fa-e657-42e4-80d3-ff5d8dd1743c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.469136] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f631d855-e3a5-4486-8a43-b1ef5bd8bf07 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.498182] env[61936]: DEBUG nova.compute.manager [req-fe112b98-05e5-40a3-98f6-eddf5d41db85 req-1bbde783-7259-47c3-8105-b258f984fd56 service nova] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Detach interface failed, port_id=39c4ceec-fef3-4d7d-8a6b-27fc061802a0, reason: Instance 628882ff-2f0f-49ae-8454-d3791fb55545 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 917.546493] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253141, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.178416} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.546881] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.548905] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7b2fe3-8982-4caf-baa7-fa778a09e7e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.577968] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/fa845f8e-957c-4c0b-a647-190c32989dcd.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.578769] env[61936]: DEBUG oslo_concurrency.lockutils [None req-24223d48-5db0-4a35-bcab-213e0298a861 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "3b09b9f5-9084-4bf4-a441-fb4b2d13c73e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.535s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.582454] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0a824d7-ab66-4d8f-8ff5-0c5e355b5b4a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.613441] env[61936]: DEBUG oslo_vmware.api [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253142, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.500052} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.616074] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.616310] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.616512] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.616688] env[61936]: INFO nova.compute.manager [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Took 1.15 seconds to destroy the instance on the hypervisor. [ 917.616930] env[61936]: DEBUG oslo.service.loopingcall [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.617200] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 917.617200] env[61936]: value = "task-1253143" [ 917.617200] env[61936]: _type = "Task" [ 917.617200] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.617804] env[61936]: DEBUG nova.compute.manager [-] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 917.617911] env[61936]: DEBUG nova.network.neutron [-] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.628144] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253143, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.891977] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b74f7c-2d44-4e14-b8ce-848823f6be13 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.900623] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4309107-db68-44c1-be45-9e6a9c1b6893 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.945047] env[61936]: INFO nova.compute.manager [-] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Took 1.34 seconds to deallocate network for instance. [ 917.948646] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92e1311-b6f0-43e9-b518-dc18ae02b5b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.957584] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdf7bbe-f7e3-40b7-a01d-fca34142748e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.972625] env[61936]: DEBUG nova.compute.provider_tree [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.129559] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253143, 'name': ReconfigVM_Task, 'duration_secs': 0.423109} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.129662] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Reconfigured VM instance instance-0000004f to attach disk [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/fa845f8e-957c-4c0b-a647-190c32989dcd.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.130381] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c7e1389-7f85-40e9-b0a5-bb718a60ccce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.138092] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 918.138092] env[61936]: value = "task-1253144" [ 918.138092] env[61936]: _type = "Task" [ 918.138092] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.154069] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253144, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.397480] env[61936]: DEBUG nova.compute.manager [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Stashing vm_state: active {{(pid=61936) _prep_resize /opt/stack/nova/nova/compute/manager.py:6015}} [ 918.478534] env[61936]: DEBUG nova.scheduler.client.report [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 918.526855] env[61936]: INFO nova.compute.manager [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Took 0.58 seconds to detach 1 volumes for instance. [ 918.650337] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253144, 'name': Rename_Task, 'duration_secs': 0.410205} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.650755] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.651125] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae47e282-e7a1-4e45-bdf9-620527229fa7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.661690] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 918.661690] env[61936]: value = "task-1253145" [ 918.661690] env[61936]: _type = "Task" [ 918.661690] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.671158] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253145, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.693667] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "2797bce9-f221-49b5-ab2b-42df3f347497" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.693946] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.694201] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "2797bce9-f221-49b5-ab2b-42df3f347497-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.694418] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.694605] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.698146] env[61936]: INFO nova.compute.manager [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Terminating instance [ 918.911813] env[61936]: DEBUG nova.network.neutron [-] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.922359] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.986035] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.961s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.990480] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.947s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.991807] env[61936]: INFO nova.compute.claims [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 918.998411] env[61936]: DEBUG nova.compute.manager [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Received event network-changed-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 918.998617] env[61936]: DEBUG nova.compute.manager [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Refreshing instance network info cache due to event network-changed-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 919.001140] env[61936]: DEBUG oslo_concurrency.lockutils [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] Acquiring lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.001349] env[61936]: DEBUG oslo_concurrency.lockutils [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] Acquired lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.002867] env[61936]: DEBUG nova.network.neutron [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Refreshing network info cache for port d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.023987] env[61936]: INFO nova.scheduler.client.report [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleted allocations for instance aa37292e-8499-46aa-8021-cb8d88b5a35a [ 919.038303] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.170372] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253145, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.204816] env[61936]: DEBUG nova.compute.manager [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 919.207479] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.209115] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46b7def-509f-47d1-8b98-89bf6a400023 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.233078] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.233956] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-433be7ee-5286-489f-bf78-46b30bc437e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.244367] env[61936]: DEBUG oslo_vmware.api [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 919.244367] env[61936]: value = "task-1253146" [ 919.244367] env[61936]: _type = "Task" [ 919.244367] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.261127] env[61936]: DEBUG oslo_vmware.api [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.395427] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.395671] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.416166] env[61936]: INFO nova.compute.manager [-] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Took 1.80 seconds to deallocate network for instance. [ 919.426823] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "a99a6fef-3cef-409a-b001-4aca97f852c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.427291] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.532442] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ea145183-1736-47a0-8edc-d90594f80f83 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "aa37292e-8499-46aa-8021-cb8d88b5a35a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.246s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.677355] env[61936]: DEBUG oslo_vmware.api [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253145, 'name': PowerOnVM_Task, 'duration_secs': 0.715975} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.677856] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.678223] env[61936]: INFO nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Took 10.20 seconds to spawn the instance on the hypervisor. [ 919.678223] env[61936]: DEBUG nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 919.683866] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db169bc5-fd22-4dd9-9001-36f96bbb3664 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.757261] env[61936]: DEBUG oslo_vmware.api [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253146, 'name': PowerOffVM_Task, 'duration_secs': 0.239255} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.757261] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 919.757420] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 919.757992] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53e3b4c8-00d9-476e-a3a1-3b34460a9478 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.806423] env[61936]: DEBUG nova.network.neutron [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updated VIF entry in instance network info cache for port d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.807038] env[61936]: DEBUG nova.network.neutron [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updating instance_info_cache with network_info: [{"id": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "address": "fa:16:3e:bb:c3:ef", "network": {"id": "0fae2486-b0c5-4d6e-8408-953cc1b06267", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2026741414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c21a56ef57b4d78bb6b1e6343efc442", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd0bdd1e2-85", "ovs_interfaceid": "d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.821773] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.822119] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.822386] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Deleting the datastore file [datastore2] 2797bce9-f221-49b5-ab2b-42df3f347497 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.822743] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cac95c2-7566-4d7e-90f3-77c93449b768 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.832418] env[61936]: DEBUG oslo_vmware.api [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 919.832418] env[61936]: value = "task-1253148" [ 919.832418] env[61936]: _type = "Task" [ 919.832418] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.847257] env[61936]: DEBUG oslo_vmware.api [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.903018] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 919.923685] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.930575] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 920.213436] env[61936]: INFO nova.compute.manager [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Took 39.09 seconds to build instance. [ 920.311206] env[61936]: DEBUG oslo_concurrency.lockutils [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] Releasing lock "refresh_cache-2797bce9-f221-49b5-ab2b-42df3f347497" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.311493] env[61936]: DEBUG nova.compute.manager [req-12d51bf2-689b-4883-8818-2d5b8a6a606b req-792d8135-c5a9-4451-a4b5-478878d42c6d service nova] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Received event network-vif-deleted-12397fd2-b04a-4019-b4fb-ca0f7b7fc152 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 920.349730] env[61936]: DEBUG oslo_vmware.api [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.425091} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.349730] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.349730] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 920.349999] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 920.349999] env[61936]: INFO nova.compute.manager [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Took 1.14 seconds to destroy the instance on the hypervisor. [ 920.350236] env[61936]: DEBUG oslo.service.loopingcall [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.350414] env[61936]: DEBUG nova.compute.manager [-] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 920.350902] env[61936]: DEBUG nova.network.neutron [-] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 920.415923] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52920ff9-d95f-405c-93c5-05c01f65a971 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.424129] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74acf2b-9c12-46de-809b-7a9f085bcc2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.428262] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.464741] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3575e4d7-6d7f-475f-a7c2-712dc1ad9d3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.473874] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c0c4a3-e301-4b27-b72b-55b1b3c6e778 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.491033] env[61936]: DEBUG nova.compute.provider_tree [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.492654] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.715702] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8b342b78-f786-4bf2-a426-d639fa14c033 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.130s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.716115] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.716348] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.716549] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.717316] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.717316] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.720336] env[61936]: INFO nova.compute.manager [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Terminating instance [ 920.994831] env[61936]: DEBUG nova.scheduler.client.report [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 921.208588] env[61936]: DEBUG nova.compute.manager [req-e765e176-763f-41fd-88cc-aecae4b7cfe1 req-24ee710e-9081-4615-9fab-adf1d6c56672 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Received event network-vif-deleted-d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 921.208787] env[61936]: INFO nova.compute.manager [req-e765e176-763f-41fd-88cc-aecae4b7cfe1 req-24ee710e-9081-4615-9fab-adf1d6c56672 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Neutron deleted interface d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f; detaching it from the instance and deleting it from the info cache [ 921.208963] env[61936]: DEBUG nova.network.neutron [req-e765e176-763f-41fd-88cc-aecae4b7cfe1 req-24ee710e-9081-4615-9fab-adf1d6c56672 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.223052] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "refresh_cache-258fed0b-ba19-4c90-981a-96bb36ba8bc8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.223365] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquired lock "refresh_cache-258fed0b-ba19-4c90-981a-96bb36ba8bc8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.223478] env[61936]: DEBUG nova.network.neutron [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.414499] env[61936]: DEBUG nova.network.neutron [-] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.501909] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.502455] env[61936]: DEBUG nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 921.505050] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.209s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.505762] env[61936]: DEBUG nova.objects.instance [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lazy-loading 'resources' on Instance uuid 18175943-35d2-4792-a8ea-e54b4ffbaf2f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.714425] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1589eef-11dc-43aa-b7c5-5037e63ca5e9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.732616] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2761b29f-9f0c-406c-9d03-39720bbcd716 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.745874] env[61936]: DEBUG nova.network.neutron [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.764432] env[61936]: DEBUG nova.compute.manager [req-e765e176-763f-41fd-88cc-aecae4b7cfe1 req-24ee710e-9081-4615-9fab-adf1d6c56672 service nova] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Detach interface failed, port_id=d0bdd1e2-85ee-4ca4-b8a5-55223a4ca12f, reason: Instance 2797bce9-f221-49b5-ab2b-42df3f347497 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 921.807871] env[61936]: DEBUG nova.network.neutron [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.813067] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.813296] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.918083] env[61936]: INFO nova.compute.manager [-] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Took 1.57 seconds to deallocate network for instance. [ 922.013409] env[61936]: DEBUG nova.compute.utils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.014923] env[61936]: DEBUG nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 922.016028] env[61936]: DEBUG nova.network.neutron [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.084208] env[61936]: DEBUG nova.policy [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9bb41ed24f244345b83f2a16f4f10804', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f1b73bc1dea4df0a995655f3bd50d14', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.313829] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Releasing lock "refresh_cache-258fed0b-ba19-4c90-981a-96bb36ba8bc8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.314765] env[61936]: DEBUG nova.compute.manager [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 922.314765] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.314900] env[61936]: DEBUG nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 922.318156] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa3b732-8dfd-48f6-9bee-8f9865e70310 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.332590] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.332590] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2a78d39-79cd-493d-a734-24f9ae60c0c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.340884] env[61936]: DEBUG oslo_vmware.api [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 922.340884] env[61936]: value = "task-1253149" [ 922.340884] env[61936]: _type = "Task" [ 922.340884] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.354813] env[61936]: DEBUG oslo_vmware.api [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.406261] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56c8db1-6be9-4486-a26d-bd3dfa343b79 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.415256] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83ae6be-d006-474a-892e-67206b8fa2bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.456213] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.461153] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57faef6f-2e0d-4f0a-8e4f-7d4c9e56f514 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.472120] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9cd4ee9-3589-413f-96a7-4d4058771afd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.487695] env[61936]: DEBUG nova.compute.provider_tree [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.518542] env[61936]: DEBUG nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 922.524516] env[61936]: DEBUG nova.network.neutron [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Successfully created port: da841dba-1be9-4d65-949a-5f8c54e14cae {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.840205] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.852558] env[61936]: DEBUG oslo_vmware.api [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253149, 'name': PowerOffVM_Task, 'duration_secs': 0.141052} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.852846] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.853029] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.853278] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f634d31-0789-4cc6-871c-760fe77421a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.878046] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.878281] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.878546] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleting the datastore file [datastore2] 258fed0b-ba19-4c90-981a-96bb36ba8bc8 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.878807] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a92631d0-2b10-49df-9468-8399da9f41b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.884819] env[61936]: DEBUG oslo_vmware.api [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for the task: (returnval){ [ 922.884819] env[61936]: value = "task-1253151" [ 922.884819] env[61936]: _type = "Task" [ 922.884819] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.892375] env[61936]: DEBUG oslo_vmware.api [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.990784] env[61936]: DEBUG nova.scheduler.client.report [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 923.395071] env[61936]: DEBUG oslo_vmware.api [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Task: {'id': task-1253151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125432} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.395332] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.395537] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.395689] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.395896] env[61936]: INFO nova.compute.manager [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Took 1.08 seconds to destroy the instance on the hypervisor. [ 923.396168] env[61936]: DEBUG oslo.service.loopingcall [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.396418] env[61936]: DEBUG nova.compute.manager [-] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 923.396473] env[61936]: DEBUG nova.network.neutron [-] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 923.420294] env[61936]: DEBUG nova.network.neutron [-] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.495418] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.497700] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.900s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.497933] env[61936]: DEBUG nova.objects.instance [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lazy-loading 'resources' on Instance uuid a8b831c7-bd6a-4218-b19e-1c43e3678c59 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.521052] env[61936]: INFO nova.scheduler.client.report [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted allocations for instance 18175943-35d2-4792-a8ea-e54b4ffbaf2f [ 923.527225] env[61936]: DEBUG nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 923.555743] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 923.556239] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.556408] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 923.557035] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.557035] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 923.557035] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 923.557206] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 923.557239] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 923.557399] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 923.558121] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 923.558416] env[61936]: DEBUG nova.virt.hardware [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 923.560435] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c7cec6-7eb0-465b-ac10-997dbbaf325c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.567425] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8215d5fd-e899-4c6b-b334-55e5302c4a83 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.922863] env[61936]: DEBUG nova.network.neutron [-] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.929137] env[61936]: DEBUG nova.compute.manager [req-045b588e-fb22-4bbc-9b5a-0eaa44cc59f3 req-dd468164-5dbd-4477-ae11-fa7d631a5ee8 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Received event network-vif-plugged-da841dba-1be9-4d65-949a-5f8c54e14cae {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 923.929374] env[61936]: DEBUG oslo_concurrency.lockutils [req-045b588e-fb22-4bbc-9b5a-0eaa44cc59f3 req-dd468164-5dbd-4477-ae11-fa7d631a5ee8 service nova] Acquiring lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.929591] env[61936]: DEBUG oslo_concurrency.lockutils [req-045b588e-fb22-4bbc-9b5a-0eaa44cc59f3 req-dd468164-5dbd-4477-ae11-fa7d631a5ee8 service nova] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.929768] env[61936]: DEBUG oslo_concurrency.lockutils [req-045b588e-fb22-4bbc-9b5a-0eaa44cc59f3 req-dd468164-5dbd-4477-ae11-fa7d631a5ee8 service nova] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.929938] env[61936]: DEBUG nova.compute.manager [req-045b588e-fb22-4bbc-9b5a-0eaa44cc59f3 req-dd468164-5dbd-4477-ae11-fa7d631a5ee8 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] No waiting events found dispatching network-vif-plugged-da841dba-1be9-4d65-949a-5f8c54e14cae {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 923.930116] env[61936]: WARNING nova.compute.manager [req-045b588e-fb22-4bbc-9b5a-0eaa44cc59f3 req-dd468164-5dbd-4477-ae11-fa7d631a5ee8 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Received unexpected event network-vif-plugged-da841dba-1be9-4d65-949a-5f8c54e14cae for instance with vm_state building and task_state spawning. [ 924.024347] env[61936]: DEBUG nova.network.neutron [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Successfully updated port: da841dba-1be9-4d65-949a-5f8c54e14cae {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.033423] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2d56439f-f1a4-42bc-b107-312f9a4c6309 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "18175943-35d2-4792-a8ea-e54b4ffbaf2f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.171s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.267099] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907c407e-ebe2-47d5-8847-54828fbf9efe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.276496] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de71e8b1-0760-4d71-bb33-f316438255ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.310532] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3469aad-6a63-423f-acaa-e6dce1de7963 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.317108] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4907b2d-1eac-4f69-b623-396584541073 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.336695] env[61936]: DEBUG nova.compute.provider_tree [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.432301] env[61936]: INFO nova.compute.manager [-] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Took 1.04 seconds to deallocate network for instance. [ 924.526917] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.527164] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.527333] env[61936]: DEBUG nova.network.neutron [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.841294] env[61936]: DEBUG nova.scheduler.client.report [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 924.938748] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.058968] env[61936]: DEBUG nova.network.neutron [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.193401] env[61936]: DEBUG nova.network.neutron [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Updating instance_info_cache with network_info: [{"id": "da841dba-1be9-4d65-949a-5f8c54e14cae", "address": "fa:16:3e:8e:e9:09", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda841dba-1b", "ovs_interfaceid": "da841dba-1be9-4d65-949a-5f8c54e14cae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.347077] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.349140] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.094s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.351114] env[61936]: INFO nova.compute.claims [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.371526] env[61936]: INFO nova.scheduler.client.report [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Deleted allocations for instance a8b831c7-bd6a-4218-b19e-1c43e3678c59 [ 925.695962] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.696239] env[61936]: DEBUG nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Instance network_info: |[{"id": "da841dba-1be9-4d65-949a-5f8c54e14cae", "address": "fa:16:3e:8e:e9:09", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda841dba-1b", "ovs_interfaceid": "da841dba-1be9-4d65-949a-5f8c54e14cae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 925.696655] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:e9:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da841dba-1be9-4d65-949a-5f8c54e14cae', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.703998] env[61936]: DEBUG oslo.service.loopingcall [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.704239] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.704467] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef4ee862-cd0c-4089-94c9-f7aff4b93d06 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.724333] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.724333] env[61936]: value = "task-1253152" [ 925.724333] env[61936]: _type = "Task" [ 925.724333] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.731853] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253152, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.878684] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a6eb8981-055d-427c-8230-beef4834a62f tempest-ServerMetadataNegativeTestJSON-659990198 tempest-ServerMetadataNegativeTestJSON-659990198-project-member] Lock "a8b831c7-bd6a-4218-b19e-1c43e3678c59" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.781s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.954083] env[61936]: DEBUG nova.compute.manager [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Received event network-changed-da841dba-1be9-4d65-949a-5f8c54e14cae {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 925.954332] env[61936]: DEBUG nova.compute.manager [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Refreshing instance network info cache due to event network-changed-da841dba-1be9-4d65-949a-5f8c54e14cae. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 925.954568] env[61936]: DEBUG oslo_concurrency.lockutils [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] Acquiring lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.954715] env[61936]: DEBUG oslo_concurrency.lockutils [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] Acquired lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.954894] env[61936]: DEBUG nova.network.neutron [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Refreshing network info cache for port da841dba-1be9-4d65-949a-5f8c54e14cae {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.235996] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253152, 'name': CreateVM_Task, 'duration_secs': 0.290305} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.236237] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 926.236967] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.237202] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.237544] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.237788] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbcc7a3f-50a3-4dda-95b8-4c3722ee8a51 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.243019] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 926.243019] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232dad4-0a6b-7681-3907-e6b67e08fce1" [ 926.243019] env[61936]: _type = "Task" [ 926.243019] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.250845] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232dad4-0a6b-7681-3907-e6b67e08fce1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.662214] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee969362-622f-488d-b6ba-2d35c027cffd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.670359] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f61faa-2b46-4f27-9bf7-d89eeabad5ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.702966] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61da5d87-9186-44db-b005-82dee1cb0466 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.713655] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180f1626-4a15-4502-9354-4af92e5699e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.727573] env[61936]: DEBUG nova.compute.provider_tree [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.753462] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232dad4-0a6b-7681-3907-e6b67e08fce1, 'name': SearchDatastore_Task, 'duration_secs': 0.009846} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.753462] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.753462] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.753646] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.753646] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.754063] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.754063] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99749f8e-808f-4345-a997-35495b6b5bd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.762919] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.763131] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.763863] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16f865b2-5817-4ee6-ad51-12b8efa0d93f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.769237] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 926.769237] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52370e9e-cc0a-1dc7-88a6-1eb44c04acb6" [ 926.769237] env[61936]: _type = "Task" [ 926.769237] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.780538] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52370e9e-cc0a-1dc7-88a6-1eb44c04acb6, 'name': SearchDatastore_Task, 'duration_secs': 0.007986} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.781308] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46d2027f-f163-4fe0-adb0-1030aa09250d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.786729] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 926.786729] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525ecc78-ceb6-18d4-33ee-dd2672464725" [ 926.786729] env[61936]: _type = "Task" [ 926.786729] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.796871] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525ecc78-ceb6-18d4-33ee-dd2672464725, 'name': SearchDatastore_Task} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.797058] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.797384] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/f3f2a0a8-cfe3-462f-a940-05580d5da32d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.797553] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a703256-7129-4458-9284-246399ad72fe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.803882] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 926.803882] env[61936]: value = "task-1253153" [ 926.803882] env[61936]: _type = "Task" [ 926.803882] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.813422] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.829819] env[61936]: DEBUG nova.network.neutron [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Updated VIF entry in instance network info cache for port da841dba-1be9-4d65-949a-5f8c54e14cae. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 926.830194] env[61936]: DEBUG nova.network.neutron [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Updating instance_info_cache with network_info: [{"id": "da841dba-1be9-4d65-949a-5f8c54e14cae", "address": "fa:16:3e:8e:e9:09", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda841dba-1b", "ovs_interfaceid": "da841dba-1be9-4d65-949a-5f8c54e14cae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.230989] env[61936]: DEBUG nova.scheduler.client.report [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 927.315201] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.422357} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.315314] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/f3f2a0a8-cfe3-462f-a940-05580d5da32d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 927.315530] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.315786] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d350775a-2fc1-4e4d-915c-b69fe294acf6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.322154] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 927.322154] env[61936]: value = "task-1253154" [ 927.322154] env[61936]: _type = "Task" [ 927.322154] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.329757] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.332356] env[61936]: DEBUG oslo_concurrency.lockutils [req-87ebbba7-e4cc-4820-8691-5b3a6fae309d req-51d11a4e-3305-4ee1-b98a-e55d947f62d0 service nova] Releasing lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.738113] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.738711] env[61936]: DEBUG nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 927.744201] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 31.650s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.744430] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.745084] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 927.745614] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.177s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.746614] env[61936]: DEBUG nova.objects.instance [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lazy-loading 'resources' on Instance uuid 3af90345-f5bf-4d25-90ac-c5389aef5c8e {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.752750] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea55722e-7192-427e-b81c-f14fc81e0440 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.762123] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926d4260-cf22-443c-b108-695bb196bbfb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.777133] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152a7673-2f3a-462e-88b5-76076d44d39d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.784104] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579dc98f-c023-4ff5-b128-5f3bb09cba4f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.821719] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180318MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 927.821875] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.832401] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074132} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.832401] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.832401] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6422ce-e306-4c5a-9e28-d6b6a005cfdf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.853556] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/f3f2a0a8-cfe3-462f-a940-05580d5da32d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.854185] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fa28a1a-ffbd-4868-9ed5-46fcefedfa1f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.873399] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 927.873399] env[61936]: value = "task-1253155" [ 927.873399] env[61936]: _type = "Task" [ 927.873399] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.882089] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253155, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.251532] env[61936]: DEBUG nova.compute.utils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.256019] env[61936]: DEBUG nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 928.256019] env[61936]: DEBUG nova.network.neutron [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 928.299130] env[61936]: DEBUG nova.policy [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b98c1bbe1044663830b18c4ce9e23d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1b7d01ffac74c6fa2306b4839041fdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 928.384408] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253155, 'name': ReconfigVM_Task, 'duration_secs': 0.323594} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.386893] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Reconfigured VM instance instance-00000050 to attach disk [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/f3f2a0a8-cfe3-462f-a940-05580d5da32d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.388113] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02d776f0-0f2d-4245-978d-9cd560e8f859 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.393998] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 928.393998] env[61936]: value = "task-1253156" [ 928.393998] env[61936]: _type = "Task" [ 928.393998] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.404947] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253156, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.524076] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b275ff-9f68-466b-b981-0ae72091ceb0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.532275] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5caf63ed-467c-4675-94d2-630b001197c8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.564419] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2c97ef-d072-403e-9a10-25b6d87afdfa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.572942] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4b96aa-7868-4ee2-a31c-d84e49ab589a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.587576] env[61936]: DEBUG nova.compute.provider_tree [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.589924] env[61936]: DEBUG nova.network.neutron [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Successfully created port: 8452d95f-2741-4d7d-a681-67321259085c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 928.758926] env[61936]: DEBUG nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 928.908942] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253156, 'name': Rename_Task, 'duration_secs': 0.268722} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.910419] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.910952] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26694dbb-cddb-4126-82d6-c9e64e0aa09e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.923019] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 928.923019] env[61936]: value = "task-1253157" [ 928.923019] env[61936]: _type = "Task" [ 928.923019] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.931178] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.093021] env[61936]: DEBUG nova.scheduler.client.report [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 929.431538] env[61936]: DEBUG oslo_vmware.api [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253157, 'name': PowerOnVM_Task, 'duration_secs': 0.493139} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.431842] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.432058] env[61936]: INFO nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Took 5.90 seconds to spawn the instance on the hypervisor. [ 929.432241] env[61936]: DEBUG nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 929.433073] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ace625-ec6d-4fae-bcdd-34d9efae7f76 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.595178] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.597901] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.002s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.598229] env[61936]: DEBUG nova.objects.instance [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lazy-loading 'resources' on Instance uuid f71581c9-b754-4ee4-bf47-ea8e52e4a2f6 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.618292] env[61936]: INFO nova.scheduler.client.report [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Deleted allocations for instance 3af90345-f5bf-4d25-90ac-c5389aef5c8e [ 929.773256] env[61936]: DEBUG nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 929.801721] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 929.802935] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.803342] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 929.803633] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.803809] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 929.803964] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 929.804202] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 929.804367] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 929.804537] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 929.804701] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 929.804872] env[61936]: DEBUG nova.virt.hardware [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 929.806085] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa920e95-4ace-443f-b107-f6fb2d94764a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.814869] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bc1fbb-eda0-4eae-8349-3f705686c4c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.954320] env[61936]: INFO nova.compute.manager [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Took 40.94 seconds to build instance. [ 930.125400] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e99584f6-84b3-4a7e-8200-dd5fb3fa1c25 tempest-ServerGroupTestJSON-1886590646 tempest-ServerGroupTestJSON-1886590646-project-member] Lock "3af90345-f5bf-4d25-90ac-c5389aef5c8e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.173s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.381799] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf77555-baff-41ff-aefd-5dd20ecabb00 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.394496] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052f9a1a-07f2-4540-a58f-320065baa8f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.427554] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa0610c-980c-43d6-a776-d41e9751ba3d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.435813] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9291814c-f437-4534-a06f-99db9a60b3e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.449617] env[61936]: DEBUG nova.compute.provider_tree [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.457903] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cc8c34ff-29b6-4135-8a5b-a54dda74cce6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.154s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.953233] env[61936]: DEBUG nova.scheduler.client.report [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 930.994538] env[61936]: DEBUG nova.network.neutron [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Successfully updated port: 8452d95f-2741-4d7d-a681-67321259085c {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.006470] env[61936]: DEBUG nova.compute.manager [req-ffdf209b-c619-430a-abb0-95e5a930c7a9 req-e225b772-f61b-4c17-b12d-37d8f80c320c service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Received event network-vif-plugged-8452d95f-2741-4d7d-a681-67321259085c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 931.006699] env[61936]: DEBUG oslo_concurrency.lockutils [req-ffdf209b-c619-430a-abb0-95e5a930c7a9 req-e225b772-f61b-4c17-b12d-37d8f80c320c service nova] Acquiring lock "b780470b-57d0-43e5-be52-539c78cada6b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.006946] env[61936]: DEBUG oslo_concurrency.lockutils [req-ffdf209b-c619-430a-abb0-95e5a930c7a9 req-e225b772-f61b-4c17-b12d-37d8f80c320c service nova] Lock "b780470b-57d0-43e5-be52-539c78cada6b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.007422] env[61936]: DEBUG oslo_concurrency.lockutils [req-ffdf209b-c619-430a-abb0-95e5a930c7a9 req-e225b772-f61b-4c17-b12d-37d8f80c320c service nova] Lock "b780470b-57d0-43e5-be52-539c78cada6b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.007684] env[61936]: DEBUG nova.compute.manager [req-ffdf209b-c619-430a-abb0-95e5a930c7a9 req-e225b772-f61b-4c17-b12d-37d8f80c320c service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] No waiting events found dispatching network-vif-plugged-8452d95f-2741-4d7d-a681-67321259085c {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 931.008148] env[61936]: WARNING nova.compute.manager [req-ffdf209b-c619-430a-abb0-95e5a930c7a9 req-e225b772-f61b-4c17-b12d-37d8f80c320c service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Received unexpected event network-vif-plugged-8452d95f-2741-4d7d-a681-67321259085c for instance with vm_state building and task_state spawning. [ 931.463381] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.862s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.466529] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.895s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.466529] env[61936]: DEBUG nova.objects.instance [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lazy-loading 'resources' on Instance uuid 9f6a1c42-e5a5-4f1d-9740-35dc6028183a {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.489789] env[61936]: INFO nova.scheduler.client.report [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleted allocations for instance f71581c9-b754-4ee4-bf47-ea8e52e4a2f6 [ 931.498671] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "refresh_cache-b780470b-57d0-43e5-be52-539c78cada6b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.498876] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "refresh_cache-b780470b-57d0-43e5-be52-539c78cada6b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.498953] env[61936]: DEBUG nova.network.neutron [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 931.579894] env[61936]: INFO nova.compute.manager [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Rescuing [ 931.580240] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.580473] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.580724] env[61936]: DEBUG nova.network.neutron [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 932.010885] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0b22a205-5a6f-4c4e-b506-685930359084 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "f71581c9-b754-4ee4-bf47-ea8e52e4a2f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.635s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.081520] env[61936]: DEBUG nova.network.neutron [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.312018] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350873d7-686c-42bb-b00d-08322f27fd66 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.318748] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967aa032-0490-4f1c-b86b-13250d23e8e7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.353020] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b11a0af-2d40-4315-b84c-23623069f1d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.361405] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7eb6b46-f3a3-4785-9261-40a7b6bd178d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.376174] env[61936]: DEBUG nova.compute.provider_tree [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.596873] env[61936]: DEBUG nova.network.neutron [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Updating instance_info_cache with network_info: [{"id": "8452d95f-2741-4d7d-a681-67321259085c", "address": "fa:16:3e:23:b1:4d", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8452d95f-27", "ovs_interfaceid": "8452d95f-2741-4d7d-a681-67321259085c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.654687] env[61936]: DEBUG nova.network.neutron [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Updating instance_info_cache with network_info: [{"id": "da841dba-1be9-4d65-949a-5f8c54e14cae", "address": "fa:16:3e:8e:e9:09", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda841dba-1b", "ovs_interfaceid": "da841dba-1be9-4d65-949a-5f8c54e14cae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.879569] env[61936]: DEBUG nova.scheduler.client.report [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 933.102328] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "refresh_cache-b780470b-57d0-43e5-be52-539c78cada6b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.102328] env[61936]: DEBUG nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Instance network_info: |[{"id": "8452d95f-2741-4d7d-a681-67321259085c", "address": "fa:16:3e:23:b1:4d", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8452d95f-27", "ovs_interfaceid": "8452d95f-2741-4d7d-a681-67321259085c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 933.102328] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:b1:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8452d95f-2741-4d7d-a681-67321259085c', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.108501] env[61936]: DEBUG oslo.service.loopingcall [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.109133] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.109595] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d27918d-f931-4f8a-b496-83141aeb7aab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.136486] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.136486] env[61936]: value = "task-1253158" [ 933.136486] env[61936]: _type = "Task" [ 933.136486] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.145529] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253158, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.157814] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-f3f2a0a8-cfe3-462f-a940-05580d5da32d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.331228] env[61936]: DEBUG nova.compute.manager [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Received event network-changed-8452d95f-2741-4d7d-a681-67321259085c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 933.331228] env[61936]: DEBUG nova.compute.manager [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Refreshing instance network info cache due to event network-changed-8452d95f-2741-4d7d-a681-67321259085c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 933.333113] env[61936]: DEBUG oslo_concurrency.lockutils [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] Acquiring lock "refresh_cache-b780470b-57d0-43e5-be52-539c78cada6b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.333440] env[61936]: DEBUG oslo_concurrency.lockutils [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] Acquired lock "refresh_cache-b780470b-57d0-43e5-be52-539c78cada6b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.335423] env[61936]: DEBUG nova.network.neutron [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Refreshing network info cache for port 8452d95f-2741-4d7d-a681-67321259085c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.385940] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.920s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.389168] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.108s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.391062] env[61936]: INFO nova.compute.claims [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.422720] env[61936]: INFO nova.scheduler.client.report [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted allocations for instance 9f6a1c42-e5a5-4f1d-9740-35dc6028183a [ 933.645877] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253158, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.695186] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.695421] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.933935] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7e9200ee-0f95-44ac-b166-54f8a99d5748 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "9f6a1c42-e5a5-4f1d-9740-35dc6028183a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.251s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.053381] env[61936]: DEBUG nova.network.neutron [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Updated VIF entry in instance network info cache for port 8452d95f-2741-4d7d-a681-67321259085c. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 934.053810] env[61936]: DEBUG nova.network.neutron [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Updating instance_info_cache with network_info: [{"id": "8452d95f-2741-4d7d-a681-67321259085c", "address": "fa:16:3e:23:b1:4d", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8452d95f-27", "ovs_interfaceid": "8452d95f-2741-4d7d-a681-67321259085c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.147956] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253158, 'name': CreateVM_Task, 'duration_secs': 0.549909} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.148202] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.148985] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.149183] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.149621] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.149944] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d073b70-0b27-47b3-9e73-5c9a99396735 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.154718] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 934.154718] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5205ee1a-d6f4-7507-466b-8d568ab88b9e" [ 934.154718] env[61936]: _type = "Task" [ 934.154718] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.164139] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5205ee1a-d6f4-7507-466b-8d568ab88b9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.197454] env[61936]: DEBUG nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 934.556881] env[61936]: DEBUG oslo_concurrency.lockutils [req-31246110-7f0c-436c-a2f9-a4301cbd13d6 req-77f69395-4370-4b16-8acb-577993c338d4 service nova] Releasing lock "refresh_cache-b780470b-57d0-43e5-be52-539c78cada6b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.667531] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5205ee1a-d6f4-7507-466b-8d568ab88b9e, 'name': SearchDatastore_Task, 'duration_secs': 0.012886} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.667833] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.668085] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.668327] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.668506] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.668694] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.668959] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8a9845f-5704-4648-96ab-b85f12d3a9de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.673721] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f51db4-8c9c-44ec-9b4b-a34132132428 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.677532] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.677734] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 934.678750] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3f36257-71c3-4fd0-8195-1b69d12f4835 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.683966] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac18cc5b-ca7e-47d3-9cf5-51b109ee9a07 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.687799] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 934.687799] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52121eca-7d91-cfdc-d8ba-843925fb5657" [ 934.687799] env[61936]: _type = "Task" [ 934.687799] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.718674] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.719580] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04f23ec3-4f9b-452b-a31f-4df3f70ddbb1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.721793] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e592f2b-1d35-4924-81e8-f82579ab8630 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.731549] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52121eca-7d91-cfdc-d8ba-843925fb5657, 'name': SearchDatastore_Task, 'duration_secs': 0.007952} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.733997] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa221bd7-1384-4cf3-a5b7-3e300c50aea1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.740727] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522ca263-f095-42d6-8c17-aff3505d9511 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.744969] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 934.744969] env[61936]: value = "task-1253159" [ 934.744969] env[61936]: _type = "Task" [ 934.744969] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.745882] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.747210] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 934.747210] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ba88aa-bdc5-6866-0201-73d6ba5e1b1d" [ 934.747210] env[61936]: _type = "Task" [ 934.747210] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.761613] env[61936]: DEBUG nova.compute.provider_tree [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.767587] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.771066] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ba88aa-bdc5-6866-0201-73d6ba5e1b1d, 'name': SearchDatastore_Task, 'duration_secs': 0.008884} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.771855] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.772148] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] b780470b-57d0-43e5-be52-539c78cada6b/b780470b-57d0-43e5-be52-539c78cada6b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 934.772408] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c366758-71a5-4f33-8898-e0031ed1933e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.779146] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 934.779146] env[61936]: value = "task-1253160" [ 934.779146] env[61936]: _type = "Task" [ 934.779146] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.787579] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253160, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.259415] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253159, 'name': PowerOffVM_Task, 'duration_secs': 0.225424} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.259788] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.260880] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ae7410-8bda-48dd-9636-7b967ed14ca6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.264203] env[61936]: DEBUG nova.scheduler.client.report [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 935.289144] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934483eb-2c63-4b89-81ac-2bb5e31f4d29 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.301800] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253160, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.321376] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.321622] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24899435-5e80-4db1-b169-2de7e5474762 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.329163] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 935.329163] env[61936]: value = "task-1253161" [ 935.329163] env[61936]: _type = "Task" [ 935.329163] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.336262] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.454878] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.455165] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.769624] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.770180] env[61936]: DEBUG nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 935.772893] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 16.851s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.799067] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253160, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523486} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.799392] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] b780470b-57d0-43e5-be52-539c78cada6b/b780470b-57d0-43e5-be52-539c78cada6b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.799624] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.799943] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9c0cf99-c190-4752-b428-e945edb02be1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.807651] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 935.807651] env[61936]: value = "task-1253162" [ 935.807651] env[61936]: _type = "Task" [ 935.807651] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.816639] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253162, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.839354] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 935.839686] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.839978] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.840191] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.840490] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.840758] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebe9467e-3245-47a1-baeb-904a2372a663 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.850262] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.850489] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.851432] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27edb83f-40a9-4463-b6f2-44d78b9901aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.856980] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 935.856980] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b87f5b-f8a1-4ef7-b11c-153d89aa6feb" [ 935.856980] env[61936]: _type = "Task" [ 935.856980] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.864327] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b87f5b-f8a1-4ef7-b11c-153d89aa6feb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.957081] env[61936]: DEBUG nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 936.276957] env[61936]: DEBUG nova.compute.utils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.282069] env[61936]: INFO nova.compute.claims [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.285940] env[61936]: DEBUG nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 936.285940] env[61936]: DEBUG nova.network.neutron [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.317122] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253162, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094557} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.317395] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.318194] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c610b6aa-60b6-48c7-bd95-e24d99ca1a38 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.340094] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] b780470b-57d0-43e5-be52-539c78cada6b/b780470b-57d0-43e5-be52-539c78cada6b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.340422] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b49dc453-7462-4193-b3b8-454e25c5e75a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.355965] env[61936]: DEBUG nova.policy [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7575c06f47234afdbe91ba9b2483ed91', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '86af7b714e214447b617777a3a8ce0fe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.363171] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 936.363171] env[61936]: value = "task-1253163" [ 936.363171] env[61936]: _type = "Task" [ 936.363171] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.366623] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b87f5b-f8a1-4ef7-b11c-153d89aa6feb, 'name': SearchDatastore_Task, 'duration_secs': 0.009773} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.369996] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c398437-f104-44d5-a6e6-a8629921d764 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.375675] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 936.375675] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525f10b2-d541-c02f-6b90-e705c304db20" [ 936.375675] env[61936]: _type = "Task" [ 936.375675] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.378471] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253163, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.387386] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525f10b2-d541-c02f-6b90-e705c304db20, 'name': SearchDatastore_Task, 'duration_secs': 0.009354} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.387386] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.387386] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk. {{(pid=61936) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 936.387682] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bcf1c464-fb2b-4b68-830d-53efa8e794b6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.396063] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 936.396063] env[61936]: value = "task-1253164" [ 936.396063] env[61936]: _type = "Task" [ 936.396063] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.405111] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253164, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.484380] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.786805] env[61936]: DEBUG nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 936.793803] env[61936]: INFO nova.compute.resource_tracker [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating resource usage from migration 43339264-3f33-4305-8859-baeb09cac628 [ 936.874771] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253163, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.915303] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253164, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447519} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.915600] env[61936]: INFO nova.virt.vmwareapi.ds_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk. [ 936.916518] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c161175-3d9e-4da2-a4cc-6d742095b502 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.948025] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.948025] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc920632-88c3-4401-8f9f-9ee9bb8afa22 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.968097] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 936.968097] env[61936]: value = "task-1253165" [ 936.968097] env[61936]: _type = "Task" [ 936.968097] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.977217] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253165, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.096817] env[61936]: DEBUG nova.network.neutron [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Successfully created port: bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.130871] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3e7e9c-dbdf-4ecf-84a3-5fa9424186ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.138605] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215449d1-0c00-443f-9561-f25d896f2397 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.177779] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbee616-c59a-4619-b63c-0d88b2ad8f0b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.185940] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93a6387-5726-4177-932a-d2df5e0c1fed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.202080] env[61936]: DEBUG nova.compute.provider_tree [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.376519] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253163, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.484641] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.705717] env[61936]: DEBUG nova.scheduler.client.report [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 937.800721] env[61936]: DEBUG nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 937.830102] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 937.830448] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.830701] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 937.831020] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.831202] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 937.831369] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 937.831594] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 937.831756] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 937.831924] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 937.832123] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 937.832317] env[61936]: DEBUG nova.virt.hardware [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 937.833248] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c51e13-967c-44ea-9507-545f0b32ba0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.841223] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf43d22-00d1-4dd0-af25-0e2623489c26 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.875072] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253163, 'name': ReconfigVM_Task, 'duration_secs': 1.289853} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.875417] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Reconfigured VM instance instance-00000051 to attach disk [datastore1] b780470b-57d0-43e5-be52-539c78cada6b/b780470b-57d0-43e5-be52-539c78cada6b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.876121] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cb3b289-eddd-4430-a3c7-9e8b9901ef05 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.882438] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 937.882438] env[61936]: value = "task-1253166" [ 937.882438] env[61936]: _type = "Task" [ 937.882438] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.890823] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253166, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.982074] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253165, 'name': ReconfigVM_Task, 'duration_secs': 0.977546} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.982426] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Reconfigured VM instance instance-00000050 to attach disk [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.983311] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd0d997-33ba-4bb1-8d33-6625eb7c4a04 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.009490] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29c8bfe0-61da-4283-80ea-fe4a675c6563 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.024441] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 938.024441] env[61936]: value = "task-1253167" [ 938.024441] env[61936]: _type = "Task" [ 938.024441] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.033747] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253167, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.212034] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.439s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.212271] env[61936]: INFO nova.compute.manager [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Migrating [ 938.212533] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.212720] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.213997] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.178s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.214241] env[61936]: DEBUG nova.objects.instance [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'resources' on Instance uuid 628882ff-2f0f-49ae-8454-d3791fb55545 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.392014] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253166, 'name': Rename_Task, 'duration_secs': 0.13739} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.392324] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.392566] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b71427a8-7566-4d39-b637-10edd5f92743 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.398586] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 938.398586] env[61936]: value = "task-1253168" [ 938.398586] env[61936]: _type = "Task" [ 938.398586] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.406031] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.535105] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253167, 'name': ReconfigVM_Task, 'duration_secs': 0.148375} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.535392] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.535650] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d76931b-f9d8-4de5-8db5-912439153d06 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.542400] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 938.542400] env[61936]: value = "task-1253169" [ 938.542400] env[61936]: _type = "Task" [ 938.542400] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.549847] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.719984] env[61936]: INFO nova.compute.rpcapi [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 938.720738] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.904827] env[61936]: DEBUG nova.compute.manager [req-55ef0419-89a7-4887-b709-918f9f1d5b07 req-fad32e92-1a7a-4dae-b44f-59684355d7fc service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Received event network-vif-plugged-bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 938.909022] env[61936]: DEBUG oslo_concurrency.lockutils [req-55ef0419-89a7-4887-b709-918f9f1d5b07 req-fad32e92-1a7a-4dae-b44f-59684355d7fc service nova] Acquiring lock "1d25e242-3542-4707-9112-3711e17df577-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.909022] env[61936]: DEBUG oslo_concurrency.lockutils [req-55ef0419-89a7-4887-b709-918f9f1d5b07 req-fad32e92-1a7a-4dae-b44f-59684355d7fc service nova] Lock "1d25e242-3542-4707-9112-3711e17df577-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.909022] env[61936]: DEBUG oslo_concurrency.lockutils [req-55ef0419-89a7-4887-b709-918f9f1d5b07 req-fad32e92-1a7a-4dae-b44f-59684355d7fc service nova] Lock "1d25e242-3542-4707-9112-3711e17df577-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.909022] env[61936]: DEBUG nova.compute.manager [req-55ef0419-89a7-4887-b709-918f9f1d5b07 req-fad32e92-1a7a-4dae-b44f-59684355d7fc service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] No waiting events found dispatching network-vif-plugged-bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 938.909022] env[61936]: WARNING nova.compute.manager [req-55ef0419-89a7-4887-b709-918f9f1d5b07 req-fad32e92-1a7a-4dae-b44f-59684355d7fc service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Received unexpected event network-vif-plugged-bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 for instance with vm_state building and task_state spawning. [ 938.912458] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253168, 'name': PowerOnVM_Task} progress is 96%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.049822] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc9e8a7-e3f5-4355-ae9f-8ef34610f52d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.063703] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ac5390-3fdf-4adc-aa25-d1058f2ad0ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.068216] env[61936]: DEBUG oslo_vmware.api [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253169, 'name': PowerOnVM_Task, 'duration_secs': 0.405696} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.068678] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.075186] env[61936]: DEBUG nova.compute.manager [None req-d8cf2745-4fdf-464e-a2b1-03f9df1474a1 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 939.076214] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fc14c1-26ab-4802-b43b-5afce8cccbac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.112766] env[61936]: DEBUG nova.network.neutron [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Successfully updated port: bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.115323] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223c5530-3855-459e-9f4e-a80bdc619637 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.128988] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae95cc69-c3cc-487f-885f-adf759c26280 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.144193] env[61936]: DEBUG nova.compute.provider_tree [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.253457] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.253711] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.253940] env[61936]: DEBUG nova.network.neutron [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.409016] env[61936]: DEBUG oslo_vmware.api [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253168, 'name': PowerOnVM_Task, 'duration_secs': 0.6741} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.409477] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.409750] env[61936]: INFO nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Took 9.64 seconds to spawn the instance on the hypervisor. [ 939.410090] env[61936]: DEBUG nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 939.410993] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add2c2a3-e3de-4b35-8562-422b4d648e50 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.619674] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "refresh_cache-1d25e242-3542-4707-9112-3711e17df577" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.622015] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquired lock "refresh_cache-1d25e242-3542-4707-9112-3711e17df577" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.622015] env[61936]: DEBUG nova.network.neutron [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.646897] env[61936]: DEBUG nova.scheduler.client.report [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 939.934960] env[61936]: INFO nova.compute.manager [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Took 44.70 seconds to build instance. [ 940.151629] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.937s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.154012] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.231s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.154600] env[61936]: DEBUG nova.objects.instance [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'resources' on Instance uuid e15be950-1c32-4801-b4e4-a4d31ff0cfec {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.175976] env[61936]: DEBUG nova.network.neutron [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance_info_cache with network_info: [{"id": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "address": "fa:16:3e:34:7d:9b", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap432e9691-4e", "ovs_interfaceid": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.182727] env[61936]: DEBUG nova.network.neutron [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.192054] env[61936]: INFO nova.scheduler.client.report [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted allocations for instance 628882ff-2f0f-49ae-8454-d3791fb55545 [ 940.413314] env[61936]: DEBUG nova.network.neutron [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Updating instance_info_cache with network_info: [{"id": "bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19", "address": "fa:16:3e:79:3c:d7", "network": {"id": "10fd3b81-80c4-4bad-813e-427e0e772068", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-569695452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86af7b714e214447b617777a3a8ce0fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdcc6520-e3", "ovs_interfaceid": "bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.435948] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0d8f358d-14c1-4e5c-b7d5-3c3937f3e324 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "b780470b-57d0-43e5-be52-539c78cada6b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.811s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.680386] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.700583] env[61936]: DEBUG oslo_concurrency.lockutils [None req-46e0548a-320c-4f9e-9819-b504f0e6cc7a tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "628882ff-2f0f-49ae-8454-d3791fb55545" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.775s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.889133] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d6fb7d-c27e-426c-a066-b420a27d46e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.898170] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52423ae-9bd7-4840-8f85-e3755db09a46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.926633] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Releasing lock "refresh_cache-1d25e242-3542-4707-9112-3711e17df577" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.926968] env[61936]: DEBUG nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Instance network_info: |[{"id": "bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19", "address": "fa:16:3e:79:3c:d7", "network": {"id": "10fd3b81-80c4-4bad-813e-427e0e772068", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-569695452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86af7b714e214447b617777a3a8ce0fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdcc6520-e3", "ovs_interfaceid": "bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 940.927829] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:3c:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7e0240aa-a694-48fc-a0f9-6f2d3e71aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.935283] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Creating folder: Project (86af7b714e214447b617777a3a8ce0fe). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 940.936029] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3428c5-fda6-4eaf-8946-32f93d90f0ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.939613] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f71e660-d368-4fd6-b583-75e022625498 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.942264] env[61936]: DEBUG nova.compute.manager [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Received event network-changed-bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 940.942451] env[61936]: DEBUG nova.compute.manager [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Refreshing instance network info cache due to event network-changed-bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 940.942663] env[61936]: DEBUG oslo_concurrency.lockutils [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] Acquiring lock "refresh_cache-1d25e242-3542-4707-9112-3711e17df577" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.942805] env[61936]: DEBUG oslo_concurrency.lockutils [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] Acquired lock "refresh_cache-1d25e242-3542-4707-9112-3711e17df577" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.943018] env[61936]: DEBUG nova.network.neutron [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Refreshing network info cache for port bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.950919] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9659149d-dbb3-4eab-809f-bc3e341a7d58 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.957532] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Created folder: Project (86af7b714e214447b617777a3a8ce0fe) in parent group-v269874. [ 940.957660] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Creating folder: Instances. Parent ref: group-v269984. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 940.958241] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cde37d3f-c30b-45b4-bab8-4e65b09ed9c6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.969756] env[61936]: DEBUG nova.compute.provider_tree [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.978054] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Created folder: Instances in parent group-v269984. [ 940.978054] env[61936]: DEBUG oslo.service.loopingcall [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.979047] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d25e242-3542-4707-9112-3711e17df577] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.979047] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3da0f90d-7466-43ea-bdb3-4ca57629199b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.998052] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.998052] env[61936]: value = "task-1253172" [ 940.998052] env[61936]: _type = "Task" [ 940.998052] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.008103] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253172, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.474901] env[61936]: DEBUG nova.scheduler.client.report [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 941.509947] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253172, 'name': CreateVM_Task, 'duration_secs': 0.296426} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.510280] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d25e242-3542-4707-9112-3711e17df577] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.511167] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.511167] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.511779] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.512060] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b8cf18e-5533-42e1-ac03-16f934f9dbae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.517315] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 941.517315] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52587dab-6c4d-75e9-aee2-fc7549d67734" [ 941.517315] env[61936]: _type = "Task" [ 941.517315] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.526475] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52587dab-6c4d-75e9-aee2-fc7549d67734, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.769032] env[61936]: DEBUG nova.compute.manager [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.770615] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ada2085-f19b-4089-8d43-53b12da842db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.773588] env[61936]: INFO nova.compute.manager [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Rescuing [ 941.773830] env[61936]: DEBUG oslo_concurrency.lockutils [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.773997] env[61936]: DEBUG oslo_concurrency.lockutils [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.774201] env[61936]: DEBUG nova.network.neutron [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.869862] env[61936]: DEBUG nova.network.neutron [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Updated VIF entry in instance network info cache for port bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.870275] env[61936]: DEBUG nova.network.neutron [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Updating instance_info_cache with network_info: [{"id": "bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19", "address": "fa:16:3e:79:3c:d7", "network": {"id": "10fd3b81-80c4-4bad-813e-427e0e772068", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-569695452-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "86af7b714e214447b617777a3a8ce0fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbdcc6520-e3", "ovs_interfaceid": "bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.980412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.984535] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.556s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.985488] env[61936]: INFO nova.compute.claims [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.998300] env[61936]: INFO nova.scheduler.client.report [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted allocations for instance e15be950-1c32-4801-b4e4-a4d31ff0cfec [ 942.029833] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52587dab-6c4d-75e9-aee2-fc7549d67734, 'name': SearchDatastore_Task, 'duration_secs': 0.00883} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.030266] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.030796] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.030796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.030922] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.031094] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.031380] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49058909-dfaa-4c44-b884-ed984e19fdf4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.041610] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.041813] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.042768] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4c3e651-14a0-4089-92cf-55e5feaaf60d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.049547] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 942.049547] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522c79a2-092c-0070-c48c-bb25a70e13ee" [ 942.049547] env[61936]: _type = "Task" [ 942.049547] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.057745] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522c79a2-092c-0070-c48c-bb25a70e13ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.197207] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe52ac4-d6bb-4d64-b6e0-c59de7c97ae2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.216008] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance '9bd97f84-7c21-44ad-9ed0-d7e4097e264e' progress to 0 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 942.286339] env[61936]: INFO nova.compute.manager [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] instance snapshotting [ 942.289259] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eece509-70b5-4a3e-9a3d-223e41ee28db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.308629] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b27322-e736-4fb8-b840-adf7a8c029dc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.373367] env[61936]: DEBUG oslo_concurrency.lockutils [req-ede0a46c-5a85-4620-abdf-7a9ab68ddc47 req-48745cd6-8178-4b11-a89c-a66e3545fb8d service nova] Releasing lock "refresh_cache-1d25e242-3542-4707-9112-3711e17df577" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.506918] env[61936]: DEBUG oslo_concurrency.lockutils [None req-10187004-10f8-44bf-b25a-81e670ec2298 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "e15be950-1c32-4801-b4e4-a4d31ff0cfec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.569s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.566022] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522c79a2-092c-0070-c48c-bb25a70e13ee, 'name': SearchDatastore_Task, 'duration_secs': 0.008303} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.566022] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd6abde5-7755-457b-a3eb-69e4f6baa535 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.570654] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 942.570654] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523ee32c-2779-fc90-d64c-36e55669a699" [ 942.570654] env[61936]: _type = "Task" [ 942.570654] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.578479] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523ee32c-2779-fc90-d64c-36e55669a699, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.634234] env[61936]: DEBUG nova.network.neutron [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Updating instance_info_cache with network_info: [{"id": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "address": "fa:16:3e:0e:6b:01", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6719762-3a", "ovs_interfaceid": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.722401] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.722780] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c3c55a3c-2f8b-4e5c-a315-0388b31e91c6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.731457] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 942.731457] env[61936]: value = "task-1253173" [ 942.731457] env[61936]: _type = "Task" [ 942.731457] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.744042] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253173, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.822903] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Creating Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 942.823287] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b1ed1afb-bb65-4b83-82bc-943fc4dec313 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.830665] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 942.830665] env[61936]: value = "task-1253174" [ 942.830665] env[61936]: _type = "Task" [ 942.830665] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.839521] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253174, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.988522] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.988764] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.080489] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523ee32c-2779-fc90-d64c-36e55669a699, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.083026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.083167] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 1d25e242-3542-4707-9112-3711e17df577/1d25e242-3542-4707-9112-3711e17df577.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.083617] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-622ea333-63d3-42f8-aebd-7a6b9874128c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.091743] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 943.091743] env[61936]: value = "task-1253175" [ 943.091743] env[61936]: _type = "Task" [ 943.091743] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.099098] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.136700] env[61936]: DEBUG oslo_concurrency.lockutils [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.244932] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253173, 'name': PowerOffVM_Task, 'duration_secs': 0.204284} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.245272] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.245550] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance '9bd97f84-7c21-44ad-9ed0-d7e4097e264e' progress to 17 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 943.283941] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719cd9e7-6724-4daa-87cf-e3e06608be0f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.295786] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63750ed8-c1cf-4491-b5b1-dcd2e1e274cb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.337868] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a222c727-a343-467d-9313-411d427a8fd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.350422] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956fb7ea-67c1-4318-a29c-6d13d57f6225 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.354697] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253174, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.367657] env[61936]: DEBUG nova.compute.provider_tree [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.494258] env[61936]: DEBUG nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 943.601856] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443187} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.602192] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 1d25e242-3542-4707-9112-3711e17df577/1d25e242-3542-4707-9112-3711e17df577.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 943.602379] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.603029] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24cf657c-c170-4841-9ec3-8aee153b4459 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.609013] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 943.609013] env[61936]: value = "task-1253176" [ 943.609013] env[61936]: _type = "Task" [ 943.609013] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.617412] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253176, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.627074] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "10669072-a8e8-4262-b390-8bf31af6fafe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.627351] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "10669072-a8e8-4262-b390-8bf31af6fafe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.627626] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "10669072-a8e8-4262-b390-8bf31af6fafe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.627845] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "10669072-a8e8-4262-b390-8bf31af6fafe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.628043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "10669072-a8e8-4262-b390-8bf31af6fafe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.630305] env[61936]: INFO nova.compute.manager [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Terminating instance [ 943.752398] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 943.752656] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.752817] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 943.753014] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.753168] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 943.753326] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 943.753520] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 943.753677] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 943.753840] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 943.755317] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 943.755641] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 943.760961] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c872becc-072b-4dc3-87a7-ad23934b3e07 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.777965] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 943.777965] env[61936]: value = "task-1253177" [ 943.777965] env[61936]: _type = "Task" [ 943.777965] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.787676] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.845531] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253174, 'name': CreateSnapshot_Task, 'duration_secs': 0.740539} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.845827] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Created Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 943.846666] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a051cc58-4974-4fc7-9e86-80468ba5c555 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.870877] env[61936]: DEBUG nova.scheduler.client.report [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 944.017108] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.119974] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065389} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.120264] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.121328] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0da14d-93a0-4faf-847b-a8224e887b1c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.135306] env[61936]: DEBUG nova.compute.manager [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 944.135519] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.144713] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 1d25e242-3542-4707-9112-3711e17df577/1d25e242-3542-4707-9112-3711e17df577.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.145669] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14862f3a-180e-4310-bb8c-a077a3169c10 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.149253] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89178868-8bae-4811-9060-b8ef76b66698 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.168987] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.171200] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-346c192c-8e29-4f46-95f3-fcf1a5a2195d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.172814] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 944.172814] env[61936]: value = "task-1253178" [ 944.172814] env[61936]: _type = "Task" [ 944.172814] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.179563] env[61936]: DEBUG oslo_vmware.api [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 944.179563] env[61936]: value = "task-1253179" [ 944.179563] env[61936]: _type = "Task" [ 944.179563] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.183368] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253178, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.191851] env[61936]: DEBUG oslo_vmware.api [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.287455] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253177, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.366475] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Creating linked-clone VM from snapshot {{(pid=61936) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 944.366475] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cbc3d25f-4a6c-4326-9024-1a3b215208cb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.374183] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 944.374183] env[61936]: value = "task-1253180" [ 944.374183] env[61936]: _type = "Task" [ 944.374183] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.382755] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.383329] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 944.385747] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253180, 'name': CloneVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.386214] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.894s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.387553] env[61936]: INFO nova.compute.claims [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.674960] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.675663] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40fcff4e-ea89-4bb3-af50-d0269f495271 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.694508] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.694897] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 944.694897] env[61936]: value = "task-1253181" [ 944.694897] env[61936]: _type = "Task" [ 944.694897] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.703736] env[61936]: DEBUG oslo_vmware.api [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253179, 'name': PowerOffVM_Task, 'duration_secs': 0.379598} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.704560] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.704814] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 944.705146] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-957eb02b-34c7-45a7-9a64-d94fd4a2f42d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.711866] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253181, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.788924] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253177, 'name': ReconfigVM_Task, 'duration_secs': 0.541555} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.789282] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance '9bd97f84-7c21-44ad-9ed0-d7e4097e264e' progress to 33 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 944.884197] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253180, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.887585] env[61936]: DEBUG nova.compute.utils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.889126] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 944.889296] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 944.895711] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 944.896811] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 944.896811] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleting the datastore file [datastore2] 10669072-a8e8-4262-b390-8bf31af6fafe {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.896811] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b1ef0f5-81af-427b-8531-bdecc99a5956 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.903368] env[61936]: DEBUG oslo_vmware.api [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 944.903368] env[61936]: value = "task-1253183" [ 944.903368] env[61936]: _type = "Task" [ 944.903368] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.912222] env[61936]: DEBUG oslo_vmware.api [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.997914] env[61936]: DEBUG nova.policy [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6aec9537a6bc46b6b55b3163f1b59cc5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91a49a1ba2c34c8080cef82a7355cd95', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 945.188189] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253178, 'name': ReconfigVM_Task, 'duration_secs': 0.592612} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.188583] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 1d25e242-3542-4707-9112-3711e17df577/1d25e242-3542-4707-9112-3711e17df577.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.189433] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6b21355-f162-4a18-8a44-9a41fec7fe46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.196679] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 945.196679] env[61936]: value = "task-1253184" [ 945.196679] env[61936]: _type = "Task" [ 945.196679] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.210948] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253181, 'name': PowerOffVM_Task, 'duration_secs': 0.494092} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.215480] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.215910] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253184, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.216930] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9095ca67-a6e5-4acb-a423-0697be39458e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.242424] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f0da78-7dc2-403c-9e7a-726e7779f33d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.275146] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.275461] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cff19b07-4be0-4bac-ad0a-7b0d4e042958 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.283035] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 945.283035] env[61936]: value = "task-1253185" [ 945.283035] env[61936]: _type = "Task" [ 945.283035] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.290876] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253185, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 945.297028] env[61936]: DEBUG nova.virt.hardware [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 945.302727] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Reconfiguring VM instance instance-0000004e to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 945.303188] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f1aab05-9eba-4857-b5b8-a71cea72ae02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.324457] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 945.324457] env[61936]: value = "task-1253186" [ 945.324457] env[61936]: _type = "Task" [ 945.324457] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.333322] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253186, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.385097] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253180, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.392072] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 945.412797] env[61936]: DEBUG oslo_vmware.api [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21242} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.413123] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.413337] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.415301] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.415301] env[61936]: INFO nova.compute.manager [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Took 1.28 seconds to destroy the instance on the hypervisor. [ 945.415301] env[61936]: DEBUG oslo.service.loopingcall [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.415301] env[61936]: DEBUG nova.compute.manager [-] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 945.415301] env[61936]: DEBUG nova.network.neutron [-] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.604773] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Successfully created port: e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.688423] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddc358e-eda5-4f20-9650-60860450af27 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.696350] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ba6ce9-8701-4a86-ad89-6b005a862320 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.732095] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be641ea1-c415-4f62-9e08-d5df58123c5b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.737902] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253184, 'name': Rename_Task, 'duration_secs': 0.280491} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.738704] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.739056] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c77982b7-e2c1-4faa-a33f-6837ccb773ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.744405] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8100107-63cf-438a-b947-8a29bcaecf1f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.749783] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 945.749783] env[61936]: value = "task-1253187" [ 945.749783] env[61936]: _type = "Task" [ 945.749783] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.761484] env[61936]: DEBUG nova.compute.provider_tree [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.768125] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253187, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.793146] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 945.793443] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.793725] env[61936]: DEBUG oslo_concurrency.lockutils [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.793892] env[61936]: DEBUG oslo_concurrency.lockutils [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.794096] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.794352] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69592692-8d3d-4509-8f9c-a60b9c9a0f54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.801917] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.802136] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.806107] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a443203e-1a31-4db8-aae3-e61d44038e53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.809512] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 945.809512] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a834fb-5c08-4f2b-b724-3d949fc29a35" [ 945.809512] env[61936]: _type = "Task" [ 945.809512] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.817272] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a834fb-5c08-4f2b-b724-3d949fc29a35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.833643] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253186, 'name': ReconfigVM_Task, 'duration_secs': 0.228003} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.834098] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Reconfigured VM instance instance-0000004e to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 945.834909] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b686ae7a-0648-4da5-9960-2f7b5317c905 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.860400] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 9bd97f84-7c21-44ad-9ed0-d7e4097e264e/9bd97f84-7c21-44ad-9ed0-d7e4097e264e.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.860805] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8380e497-b41a-4dca-ab22-c7de1eac1059 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.887791] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253180, 'name': CloneVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.891124] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 945.891124] env[61936]: value = "task-1253188" [ 945.891124] env[61936]: _type = "Task" [ 945.891124] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.903219] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253188, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.970192] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.970660] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "0137305a-dace-4eda-9d90-7233c001176f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.261126] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253187, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.265301] env[61936]: DEBUG nova.scheduler.client.report [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.321706] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a834fb-5c08-4f2b-b724-3d949fc29a35, 'name': SearchDatastore_Task, 'duration_secs': 0.021005} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.322804] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf459765-48c6-4099-9e4f-9cee983c2747 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.329321] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 946.329321] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5206b3fa-114d-2749-21c0-cad79ad021b3" [ 946.329321] env[61936]: _type = "Task" [ 946.329321] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.339394] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5206b3fa-114d-2749-21c0-cad79ad021b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.389724] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253180, 'name': CloneVM_Task, 'duration_secs': 1.529244} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.390246] env[61936]: INFO nova.virt.vmwareapi.vmops [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Created linked-clone VM from snapshot [ 946.394266] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d55961-0a50-4a07-88aa-fc14a0808900 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.402600] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.410808] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 946.416777] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Uploading image af56f59f-c70a-4b04-91e5-d2e24b68120b {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 946.446567] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 946.446905] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.447313] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 946.448017] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.448017] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 946.448017] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 946.448156] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 946.448278] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 946.448654] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 946.449044] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 946.449372] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 946.451680] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f2f49d-4d80-41b3-99c6-5fdaf523300c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.462159] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db214acc-3b35-4631-8bce-d8c9cadb8d47 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.467991] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 946.467991] env[61936]: value = "vm-269988" [ 946.467991] env[61936]: _type = "VirtualMachine" [ 946.467991] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 946.468360] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-22c6dbf4-127b-4de4-a6df-2f121226174c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.482367] env[61936]: DEBUG nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 946.486853] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lease: (returnval){ [ 946.486853] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529991f0-1ddc-bfda-161e-8a5059ef1b25" [ 946.486853] env[61936]: _type = "HttpNfcLease" [ 946.486853] env[61936]: } obtained for exporting VM: (result){ [ 946.486853] env[61936]: value = "vm-269988" [ 946.486853] env[61936]: _type = "VirtualMachine" [ 946.486853] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 946.487288] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the lease: (returnval){ [ 946.487288] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529991f0-1ddc-bfda-161e-8a5059ef1b25" [ 946.487288] env[61936]: _type = "HttpNfcLease" [ 946.487288] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 946.494347] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 946.494347] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529991f0-1ddc-bfda-161e-8a5059ef1b25" [ 946.494347] env[61936]: _type = "HttpNfcLease" [ 946.494347] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 946.649168] env[61936]: DEBUG nova.compute.manager [req-3e545aa3-9c0a-457f-8a0d-21debc51ab29 req-a28275d8-9908-439c-8540-002feb68151c service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Received event network-vif-deleted-894b1eac-a089-4cf8-adbc-f2e646971faf {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 946.649168] env[61936]: INFO nova.compute.manager [req-3e545aa3-9c0a-457f-8a0d-21debc51ab29 req-a28275d8-9908-439c-8540-002feb68151c service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Neutron deleted interface 894b1eac-a089-4cf8-adbc-f2e646971faf; detaching it from the instance and deleting it from the info cache [ 946.649408] env[61936]: DEBUG nova.network.neutron [req-3e545aa3-9c0a-457f-8a0d-21debc51ab29 req-a28275d8-9908-439c-8540-002feb68151c service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.761780] env[61936]: DEBUG oslo_vmware.api [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253187, 'name': PowerOnVM_Task, 'duration_secs': 0.543786} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.762158] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.762266] env[61936]: INFO nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Took 8.96 seconds to spawn the instance on the hypervisor. [ 946.762446] env[61936]: DEBUG nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.763260] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5edd89-0f93-44b3-8d0e-1db821ec4e27 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.772803] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.773489] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 946.779902] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.320s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.780167] env[61936]: DEBUG nova.objects.instance [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lazy-loading 'resources' on Instance uuid 2797bce9-f221-49b5-ab2b-42df3f347497 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.841642] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5206b3fa-114d-2749-21c0-cad79ad021b3, 'name': SearchDatastore_Task, 'duration_secs': 0.013121} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.841934] env[61936]: DEBUG oslo_concurrency.lockutils [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.842227] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk. {{(pid=61936) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 946.842499] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dac7e53e-1818-4e1b-a80f-618c4c3c0f3d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.852334] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 946.852334] env[61936]: value = "task-1253190" [ 946.852334] env[61936]: _type = "Task" [ 946.852334] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.860641] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.900632] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253188, 'name': ReconfigVM_Task, 'duration_secs': 0.603809} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.900943] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 9bd97f84-7c21-44ad-9ed0-d7e4097e264e/9bd97f84-7c21-44ad-9ed0-d7e4097e264e.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.901237] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance '9bd97f84-7c21-44ad-9ed0-d7e4097e264e' progress to 50 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.000794] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 947.000794] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529991f0-1ddc-bfda-161e-8a5059ef1b25" [ 947.000794] env[61936]: _type = "HttpNfcLease" [ 947.000794] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 947.001066] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 947.001066] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529991f0-1ddc-bfda-161e-8a5059ef1b25" [ 947.001066] env[61936]: _type = "HttpNfcLease" [ 947.001066] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 947.001713] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692ab4d8-4009-47ca-a2e7-9ab1db3796fe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.009319] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525edfbc-d030-3015-a7b7-3bfc41a9a025/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 947.009519] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525edfbc-d030-3015-a7b7-3bfc41a9a025/disk-0.vmdk for reading. {{(pid=61936) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 947.011912] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.087616] env[61936]: DEBUG nova.network.neutron [-] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.109221] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b389b505-c4aa-48be-880c-530d45219a05 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.152974] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae515268-4fa9-44a5-b3af-79c210bf0b08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.167549] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7628c6ef-515e-4f2e-a4bc-0e1e4063b931 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.201927] env[61936]: DEBUG nova.compute.manager [req-3e545aa3-9c0a-457f-8a0d-21debc51ab29 req-a28275d8-9908-439c-8540-002feb68151c service nova] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Detach interface failed, port_id=894b1eac-a089-4cf8-adbc-f2e646971faf, reason: Instance 10669072-a8e8-4262-b390-8bf31af6fafe could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 947.282509] env[61936]: DEBUG nova.compute.utils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.284194] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 947.284404] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.291622] env[61936]: INFO nova.compute.manager [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Took 30.03 seconds to build instance. [ 947.364667] env[61936]: DEBUG nova.policy [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6aec9537a6bc46b6b55b3163f1b59cc5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91a49a1ba2c34c8080cef82a7355cd95', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 947.381078] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253190, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.408898] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8310d08-6e10-4777-89e5-b0be41a03c9d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.436597] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57c9e0e-b409-48de-9f5f-088966f926c4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.455513] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance '9bd97f84-7c21-44ad-9ed0-d7e4097e264e' progress to 67 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.592126] env[61936]: INFO nova.compute.manager [-] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Took 2.18 seconds to deallocate network for instance. [ 947.657596] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace54db5-1327-4c8e-92cf-d6a8794263a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.667899] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7095f2-ee92-4304-874b-264236cdb7b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.707522] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb7f496-dac3-4951-820a-05bc35481151 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.715705] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6995e92-f663-40ef-886f-c9caa1d5cc4c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.731710] env[61936]: DEBUG nova.compute.provider_tree [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.789195] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 947.793988] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a2899c18-d0ef-4187-b599-8aa219207f11 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "1d25e242-3542-4707-9112-3711e17df577" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.545s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.864496] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609048} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.865355] env[61936]: INFO nova.virt.vmwareapi.ds_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk. [ 947.866542] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14e4401-3d12-4560-abf4-70d0b50d5f2d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.895959] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.896348] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4672392d-a672-4f78-b154-abece095d97a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.918181] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 947.918181] env[61936]: value = "task-1253191" [ 947.918181] env[61936]: _type = "Task" [ 947.918181] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.927548] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253191, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.083056] env[61936]: DEBUG nova.network.neutron [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Port 432e9691-4e6f-4550-a7f4-7396d05e96d7 binding to destination host cpu-1 is already ACTIVE {{(pid=61936) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 948.100747] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.170237] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Successfully created port: c29e5d63-c97e-4401-85ae-d7f0d13b72f0 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.236458] env[61936]: DEBUG nova.scheduler.client.report [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.258688] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Successfully updated port: e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.429583] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.727881] env[61936]: DEBUG nova.compute.manager [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Received event network-vif-plugged-e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 948.728441] env[61936]: DEBUG oslo_concurrency.lockutils [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] Acquiring lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.728590] env[61936]: DEBUG oslo_concurrency.lockutils [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.728797] env[61936]: DEBUG oslo_concurrency.lockutils [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.729048] env[61936]: DEBUG nova.compute.manager [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] No waiting events found dispatching network-vif-plugged-e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 948.729276] env[61936]: WARNING nova.compute.manager [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Received unexpected event network-vif-plugged-e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 for instance with vm_state building and task_state spawning. [ 948.729485] env[61936]: DEBUG nova.compute.manager [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Received event network-changed-e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 948.729745] env[61936]: DEBUG nova.compute.manager [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Refreshing instance network info cache due to event network-changed-e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 948.729975] env[61936]: DEBUG oslo_concurrency.lockutils [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] Acquiring lock "refresh_cache-45fd1d4d-7125-43a7-afe7-53f737fb9f8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.730177] env[61936]: DEBUG oslo_concurrency.lockutils [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] Acquired lock "refresh_cache-45fd1d4d-7125-43a7-afe7-53f737fb9f8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.730409] env[61936]: DEBUG nova.network.neutron [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Refreshing network info cache for port e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 948.742586] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.966s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.745770] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.906s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.748356] env[61936]: INFO nova.compute.claims [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 948.763378] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "refresh_cache-45fd1d4d-7125-43a7-afe7-53f737fb9f8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.777963] env[61936]: INFO nova.scheduler.client.report [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Deleted allocations for instance 2797bce9-f221-49b5-ab2b-42df3f347497 [ 948.802843] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 948.831049] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 948.831332] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 948.832791] env[61936]: DEBUG nova.virt.hardware [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 948.833760] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be611bfd-3ff6-4ded-9616-4c037669c7b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.843566] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c1e003-471b-44e3-a932-22e1792b7989 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.936973] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253191, 'name': ReconfigVM_Task, 'duration_secs': 0.637258} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.937530] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Reconfigured VM instance instance-0000004f to attach disk [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.938886] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0610ad75-8a06-4be7-a397-8450c359a6c6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.966731] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-126c7c0f-988a-41b0-b653-ab2fa0c10b99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.982895] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 948.982895] env[61936]: value = "task-1253192" [ 948.982895] env[61936]: _type = "Task" [ 948.982895] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.991820] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253192, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.107955] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.107955] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.108281] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.292184] env[61936]: DEBUG nova.network.neutron [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 949.295064] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad1ca5ba-8268-46e1-b2a0-d2762decdf52 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "2797bce9-f221-49b5-ab2b-42df3f347497" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.601s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.466423] env[61936]: DEBUG nova.network.neutron [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.493915] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253192, 'name': ReconfigVM_Task, 'duration_secs': 0.21538} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.494229] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.494489] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25a57c54-2b11-44fe-934e-8619901bac3e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.502486] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 949.502486] env[61936]: value = "task-1253193" [ 949.502486] env[61936]: _type = "Task" [ 949.502486] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.510122] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.968809] env[61936]: DEBUG oslo_concurrency.lockutils [req-106ab395-653a-4a98-a7ec-8081c6e110a9 req-8e4e6c84-127e-499d-9add-5d538aa62c93 service nova] Releasing lock "refresh_cache-45fd1d4d-7125-43a7-afe7-53f737fb9f8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.969262] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "refresh_cache-45fd1d4d-7125-43a7-afe7-53f737fb9f8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.969344] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.013938] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253193, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.096340] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f988c4-b91d-42f2-9ab1-7918d4e5c2a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.104141] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c5b013-1940-4ca1-b77b-60715af953ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.141144] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f46d6e0-098f-4a96-ba21-d3776f557a14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.149539] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4361fc9-813d-427d-b7fa-4b7caaa41569 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.164617] env[61936]: DEBUG nova.compute.provider_tree [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.234624] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.234814] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.235008] env[61936]: DEBUG nova.network.neutron [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.513918] env[61936]: DEBUG oslo_vmware.api [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253193, 'name': PowerOnVM_Task, 'duration_secs': 0.519287} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.514587] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 950.518095] env[61936]: DEBUG nova.compute.manager [None req-117ece9d-278e-4e66-85ef-0e966e2b93f6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 950.522427] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a622ca2-ab82-49eb-aa42-e3626d484728 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.539242] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Successfully updated port: c29e5d63-c97e-4401-85ae-d7f0d13b72f0 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.542484] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 950.667904] env[61936]: DEBUG nova.scheduler.client.report [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 950.801413] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Updating instance_info_cache with network_info: [{"id": "e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53", "address": "fa:16:3e:b9:f8:d3", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape72c28ed-f0", "ovs_interfaceid": "e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.036751] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "refresh_cache-a99a6fef-3cef-409a-b001-4aca97f852c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.037094] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "refresh_cache-a99a6fef-3cef-409a-b001-4aca97f852c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.037286] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.136280] env[61936]: DEBUG nova.network.neutron [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance_info_cache with network_info: [{"id": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "address": "fa:16:3e:34:7d:9b", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap432e9691-4e", "ovs_interfaceid": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.164245] env[61936]: DEBUG nova.compute.manager [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Received event network-vif-plugged-c29e5d63-c97e-4401-85ae-d7f0d13b72f0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 951.164577] env[61936]: DEBUG oslo_concurrency.lockutils [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] Acquiring lock "a99a6fef-3cef-409a-b001-4aca97f852c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.164671] env[61936]: DEBUG oslo_concurrency.lockutils [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.164867] env[61936]: DEBUG oslo_concurrency.lockutils [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.165011] env[61936]: DEBUG nova.compute.manager [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] No waiting events found dispatching network-vif-plugged-c29e5d63-c97e-4401-85ae-d7f0d13b72f0 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 951.165192] env[61936]: WARNING nova.compute.manager [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Received unexpected event network-vif-plugged-c29e5d63-c97e-4401-85ae-d7f0d13b72f0 for instance with vm_state building and task_state spawning. [ 951.165354] env[61936]: DEBUG nova.compute.manager [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Received event network-changed-c29e5d63-c97e-4401-85ae-d7f0d13b72f0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 951.165586] env[61936]: DEBUG nova.compute.manager [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Refreshing instance network info cache due to event network-changed-c29e5d63-c97e-4401-85ae-d7f0d13b72f0. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 951.165739] env[61936]: DEBUG oslo_concurrency.lockutils [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] Acquiring lock "refresh_cache-a99a6fef-3cef-409a-b001-4aca97f852c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.176046] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.176226] env[61936]: DEBUG nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 951.180800] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.241s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.181212] env[61936]: DEBUG nova.objects.instance [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lazy-loading 'resources' on Instance uuid 258fed0b-ba19-4c90-981a-96bb36ba8bc8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.304218] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "refresh_cache-45fd1d4d-7125-43a7-afe7-53f737fb9f8e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.304579] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Instance network_info: |[{"id": "e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53", "address": "fa:16:3e:b9:f8:d3", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape72c28ed-f0", "ovs_interfaceid": "e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 951.305068] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:f8:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73e099e8-2acc-4628-a60d-0b4afa46b39d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.312821] env[61936]: DEBUG oslo.service.loopingcall [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.313097] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.313337] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11986d6c-eb68-4728-8857-3dbf49ab4967 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.334923] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.334923] env[61936]: value = "task-1253194" [ 951.334923] env[61936]: _type = "Task" [ 951.334923] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.345222] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253194, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.422605] env[61936]: INFO nova.compute.manager [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Rebuilding instance [ 951.470777] env[61936]: DEBUG nova.compute.manager [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 951.471684] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c101604-3b21-4bb4-9a9f-5d50c0d2d1cb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.586275] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.640750] env[61936]: DEBUG oslo_concurrency.lockutils [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.686040] env[61936]: DEBUG nova.compute.utils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 951.690320] env[61936]: DEBUG nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 951.690589] env[61936]: DEBUG nova.network.neutron [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.728821] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "1d25e242-3542-4707-9112-3711e17df577" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.729159] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "1d25e242-3542-4707-9112-3711e17df577" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.729384] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "1d25e242-3542-4707-9112-3711e17df577-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.729616] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "1d25e242-3542-4707-9112-3711e17df577-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.729797] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "1d25e242-3542-4707-9112-3711e17df577-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.732276] env[61936]: INFO nova.compute.manager [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Terminating instance [ 951.813909] env[61936]: DEBUG nova.policy [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e0323c8cb7c4081ac62a7193e209d7b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '905821a1ff2b4011994c0d2d8bc08b13', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.847031] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253194, 'name': CreateVM_Task, 'duration_secs': 0.467439} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.847233] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 951.848275] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.848502] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.849060] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.849412] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0780aa42-9942-4ef3-9e80-59a1d4462baa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.855268] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 951.855268] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ac6cf9-d3a3-4d85-a997-9e2c39096044" [ 951.855268] env[61936]: _type = "Task" [ 951.855268] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.868482] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ac6cf9-d3a3-4d85-a997-9e2c39096044, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.975056] env[61936]: DEBUG nova.network.neutron [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Updating instance_info_cache with network_info: [{"id": "c29e5d63-c97e-4401-85ae-d7f0d13b72f0", "address": "fa:16:3e:ca:63:0a", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc29e5d63-c9", "ovs_interfaceid": "c29e5d63-c97e-4401-85ae-d7f0d13b72f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.009316] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc37684-2b46-434b-bbd1-b7ffaff8ca84 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.018211] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5764a198-a155-4e03-aa5a-886e22029075 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.053605] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fdc416-2c0d-4b1d-b57c-88a58b8bb986 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.062037] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8bd91f-c65d-470c-9b87-3bd1b2f2f942 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.085148] env[61936]: DEBUG nova.compute.provider_tree [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.164153] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c315c4d-8fc7-4904-b294-b481e4fa2f9d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.183114] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59ab1ff-25ea-4c07-80eb-2a5a7084bcf3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.190449] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance '9bd97f84-7c21-44ad-9ed0-d7e4097e264e' progress to 83 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 952.197411] env[61936]: DEBUG nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 952.242078] env[61936]: DEBUG nova.compute.manager [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 952.242753] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.243888] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a7202e-fe04-4e69-a44e-20b9f0dbf5c9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.255112] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.255112] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52627839-d198-4d3e-ba78-b60d83beed31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.258676] env[61936]: DEBUG oslo_vmware.api [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 952.258676] env[61936]: value = "task-1253195" [ 952.258676] env[61936]: _type = "Task" [ 952.258676] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.269792] env[61936]: DEBUG oslo_vmware.api [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.369281] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ac6cf9-d3a3-4d85-a997-9e2c39096044, 'name': SearchDatastore_Task, 'duration_secs': 0.013689} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.369605] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.370086] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.370306] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.370489] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.370718] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.370997] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-060ab103-a2ea-479a-9fed-603164cdac87 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.379764] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.379973] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 952.380732] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6056a1b2-1d50-4d1f-b0af-c902d0ba7c4b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.386461] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 952.386461] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5202b21c-c711-8edb-343d-44f72d7623cd" [ 952.386461] env[61936]: _type = "Task" [ 952.386461] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.394961] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5202b21c-c711-8edb-343d-44f72d7623cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.458596] env[61936]: DEBUG nova.network.neutron [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Successfully created port: d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 952.480371] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "refresh_cache-a99a6fef-3cef-409a-b001-4aca97f852c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.480371] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Instance network_info: |[{"id": "c29e5d63-c97e-4401-85ae-d7f0d13b72f0", "address": "fa:16:3e:ca:63:0a", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc29e5d63-c9", "ovs_interfaceid": "c29e5d63-c97e-4401-85ae-d7f0d13b72f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 952.480371] env[61936]: DEBUG oslo_concurrency.lockutils [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] Acquired lock "refresh_cache-a99a6fef-3cef-409a-b001-4aca97f852c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.480371] env[61936]: DEBUG nova.network.neutron [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Refreshing network info cache for port c29e5d63-c97e-4401-85ae-d7f0d13b72f0 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.480371] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:63:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73e099e8-2acc-4628-a60d-0b4afa46b39d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c29e5d63-c97e-4401-85ae-d7f0d13b72f0', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.489413] env[61936]: DEBUG oslo.service.loopingcall [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.489413] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.489413] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.489413] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2b7cc02-a9fd-4b06-adea-82447b993aa4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.504104] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f9205153-ae01-4463-97f4-e8712487c6db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.511952] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 952.511952] env[61936]: value = "task-1253196" [ 952.511952] env[61936]: _type = "Task" [ 952.511952] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.513459] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.513459] env[61936]: value = "task-1253197" [ 952.513459] env[61936]: _type = "Task" [ 952.513459] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.526054] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253196, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.529889] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253197, 'name': CreateVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.575403] env[61936]: INFO nova.compute.manager [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Unrescuing [ 952.575871] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.576143] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.576345] env[61936]: DEBUG nova.network.neutron [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.587799] env[61936]: DEBUG nova.scheduler.client.report [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 952.700170] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.704468] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7981a324-e3e9-434f-aa9a-7112d1082beb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.712535] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 952.712535] env[61936]: value = "task-1253198" [ 952.712535] env[61936]: _type = "Task" [ 952.712535] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.721950] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253198, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.768422] env[61936]: DEBUG oslo_vmware.api [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253195, 'name': PowerOffVM_Task, 'duration_secs': 0.218708} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.842766] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 952.842766] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 952.842766] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8277511-1c63-453b-afca-74dcf218cda8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.842766] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 952.842766] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 952.842766] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Deleting the datastore file [datastore2] 1d25e242-3542-4707-9112-3711e17df577 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.842766] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d2e72cb-f9e6-4ac3-b012-aa8ec8b8913d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.842766] env[61936]: DEBUG oslo_vmware.api [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for the task: (returnval){ [ 952.842766] env[61936]: value = "task-1253200" [ 952.842766] env[61936]: _type = "Task" [ 952.842766] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.849775] env[61936]: DEBUG oslo_vmware.api [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.897712] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5202b21c-c711-8edb-343d-44f72d7623cd, 'name': SearchDatastore_Task, 'duration_secs': 0.010273} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.898332] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6476939f-5d14-4e11-bc2c-68e53374b1fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.904091] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 952.904091] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f86ab9-fc08-43b8-bca1-c48b7bfd41d4" [ 952.904091] env[61936]: _type = "Task" [ 952.904091] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.914210] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f86ab9-fc08-43b8-bca1-c48b7bfd41d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.031484] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253196, 'name': PowerOffVM_Task, 'duration_secs': 0.182435} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.035998] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.036875] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.037156] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253197, 'name': CreateVM_Task, 'duration_secs': 0.46414} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.039665] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ed56d63-ef02-474a-b2a2-c97f217e95a5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.041415] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.042133] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.042308] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.042618] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.043875] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16cf81f7-60ba-4e8c-a247-062409076ca9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.048252] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 953.048252] env[61936]: value = "task-1253201" [ 953.048252] env[61936]: _type = "Task" [ 953.048252] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.049780] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 953.049780] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522c2c43-8a40-e23f-db23-1c0054e10c0b" [ 953.049780] env[61936]: _type = "Task" [ 953.049780] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.062935] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522c2c43-8a40-e23f-db23-1c0054e10c0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.068343] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 953.068887] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 953.069145] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269950', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'name': 'volume-612c642f-abe9-4886-ac2f-baadfe0dcb78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a12fd59-bacd-42b3-99c1-8910f349e2cb', 'attached_at': '', 'detached_at': '', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'serial': '612c642f-abe9-4886-ac2f-baadfe0dcb78'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 953.070186] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8d95a9-a3a7-44a8-b03a-0cdb2ac9faa1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.094470] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.915s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.097504] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3fb746b-40fb-419c-ae21-8297789c539d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.100324] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 25.278s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.106777] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7c0415-4bb6-41b2-a5aa-1368283326f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.125294] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93859dda-1fad-4d6e-af6b-517e2ea792df {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.129104] env[61936]: INFO nova.scheduler.client.report [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Deleted allocations for instance 258fed0b-ba19-4c90-981a-96bb36ba8bc8 [ 953.145388] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] The volume has not been displaced from its original location: [datastore1] volume-612c642f-abe9-4886-ac2f-baadfe0dcb78/volume-612c642f-abe9-4886-ac2f-baadfe0dcb78.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 953.150946] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Reconfiguring VM instance instance-0000004d to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 953.151585] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acd851e5-07da-448f-89e7-6100588368bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.174413] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "f2844173-4114-4158-99ad-ac651efbb607" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.174664] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "f2844173-4114-4158-99ad-ac651efbb607" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.174867] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "f2844173-4114-4158-99ad-ac651efbb607-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.175092] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "f2844173-4114-4158-99ad-ac651efbb607-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.175308] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "f2844173-4114-4158-99ad-ac651efbb607-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.178735] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 953.178735] env[61936]: value = "task-1253202" [ 953.178735] env[61936]: _type = "Task" [ 953.178735] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.179188] env[61936]: INFO nova.compute.manager [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Terminating instance [ 953.190793] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.209287] env[61936]: DEBUG nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 953.225158] env[61936]: DEBUG oslo_vmware.api [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253198, 'name': PowerOnVM_Task, 'duration_secs': 0.39991} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.225481] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.225734] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-16e79bfc-b5fd-4efe-8e7b-0c2e56ca8c00 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance '9bd97f84-7c21-44ad-9ed0-d7e4097e264e' progress to 100 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 953.330114] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 953.330405] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.330571] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 953.331026] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.331026] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 953.331154] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 953.331358] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 953.331430] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 953.331579] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 953.331740] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 953.331912] env[61936]: DEBUG nova.virt.hardware [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 953.333285] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71024852-8bf0-411e-85cc-a86e73c027bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.345000] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e76e9a4-7a4b-41b2-8375-676e0741de85 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.354670] env[61936]: DEBUG oslo_vmware.api [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Task: {'id': task-1253200, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198951} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.364870] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.365092] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.365275] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.365453] env[61936]: INFO nova.compute.manager [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] [instance: 1d25e242-3542-4707-9112-3711e17df577] Took 1.12 seconds to destroy the instance on the hypervisor. [ 953.365702] env[61936]: DEBUG oslo.service.loopingcall [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.368422] env[61936]: DEBUG nova.compute.manager [-] [instance: 1d25e242-3542-4707-9112-3711e17df577] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 953.368536] env[61936]: DEBUG nova.network.neutron [-] [instance: 1d25e242-3542-4707-9112-3711e17df577] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.414066] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f86ab9-fc08-43b8-bca1-c48b7bfd41d4, 'name': SearchDatastore_Task, 'duration_secs': 0.022616} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.414347] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.414607] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 45fd1d4d-7125-43a7-afe7-53f737fb9f8e/45fd1d4d-7125-43a7-afe7-53f737fb9f8e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 953.414864] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6305a155-7dde-44b2-8764-6741cb2b4020 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.421780] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 953.421780] env[61936]: value = "task-1253203" [ 953.421780] env[61936]: _type = "Task" [ 953.421780] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.429574] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253203, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.504665] env[61936]: DEBUG nova.network.neutron [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Updated VIF entry in instance network info cache for port c29e5d63-c97e-4401-85ae-d7f0d13b72f0. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.505366] env[61936]: DEBUG nova.network.neutron [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Updating instance_info_cache with network_info: [{"id": "c29e5d63-c97e-4401-85ae-d7f0d13b72f0", "address": "fa:16:3e:ca:63:0a", "network": {"id": "32c5177d-3704-407d-a8fd-2dd315bed6f1", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-130695894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91a49a1ba2c34c8080cef82a7355cd95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73e099e8-2acc-4628-a60d-0b4afa46b39d", "external-id": "nsx-vlan-transportzone-767", "segmentation_id": 767, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc29e5d63-c9", "ovs_interfaceid": "c29e5d63-c97e-4401-85ae-d7f0d13b72f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.566135] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522c2c43-8a40-e23f-db23-1c0054e10c0b, 'name': SearchDatastore_Task, 'duration_secs': 0.012345} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.566554] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.566791] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.567039] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.567195] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.567411] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.567695] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99c728c0-50b9-4554-bc3c-b16384ff09cb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.585061] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.585311] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.586121] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cd99839-0d10-4bd5-bfd9-ae0544ad9297 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.593220] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 953.593220] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524e38d9-4808-c77a-95d4-4963fe757ad6" [ 953.593220] env[61936]: _type = "Task" [ 953.593220] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.613161] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524e38d9-4808-c77a-95d4-4963fe757ad6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.643522] env[61936]: DEBUG nova.network.neutron [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Updating instance_info_cache with network_info: [{"id": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "address": "fa:16:3e:0e:6b:01", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6719762-3a", "ovs_interfaceid": "a6719762-3a94-4d4a-9fea-ff09306d59b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.659903] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7b5295ff-6270-4109-8462-8b1f917009e9 tempest-ServerShowV247Test-778573357 tempest-ServerShowV247Test-778573357-project-member] Lock "258fed0b-ba19-4c90-981a-96bb36ba8bc8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.939s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.686556] env[61936]: DEBUG nova.compute.manager [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 953.686850] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.687670] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12caec08-480e-4cc9-8129-3e7a5097e1a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.700112] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.700457] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253202, 'name': ReconfigVM_Task, 'duration_secs': 0.272591} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.700664] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fee0837-0ea6-434d-9ab5-11e4fe9dbd6d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.702739] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Reconfigured VM instance instance-0000004d to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.706964] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30a332dd-0b9a-4410-8d6b-99eee8c020ff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.723710] env[61936]: DEBUG oslo_vmware.api [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 953.723710] env[61936]: value = "task-1253204" [ 953.723710] env[61936]: _type = "Task" [ 953.723710] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.725414] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 953.725414] env[61936]: value = "task-1253205" [ 953.725414] env[61936]: _type = "Task" [ 953.725414] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.744116] env[61936]: DEBUG oslo_vmware.api [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253204, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.747992] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253205, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.814182] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.815281] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.815500] env[61936]: INFO nova.compute.manager [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Shelving [ 953.943397] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253203, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.008559] env[61936]: DEBUG nova.compute.manager [req-7db5ee1c-f9cf-4fd2-99f4-f69bfa83f6d3 req-f6a5a8de-4463-4e12-bc14-0e9fbcd734b9 service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Received event network-vif-deleted-bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 954.009039] env[61936]: INFO nova.compute.manager [req-7db5ee1c-f9cf-4fd2-99f4-f69bfa83f6d3 req-f6a5a8de-4463-4e12-bc14-0e9fbcd734b9 service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Neutron deleted interface bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19; detaching it from the instance and deleting it from the info cache [ 954.009756] env[61936]: DEBUG nova.network.neutron [req-7db5ee1c-f9cf-4fd2-99f4-f69bfa83f6d3 req-f6a5a8de-4463-4e12-bc14-0e9fbcd734b9 service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.013125] env[61936]: DEBUG oslo_concurrency.lockutils [req-ddd3673d-cc8f-4a18-be2f-2a9345339d19 req-3030fbf6-204b-40f1-b5f5-555bbb073c1c service nova] Releasing lock "refresh_cache-a99a6fef-3cef-409a-b001-4aca97f852c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.106256] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524e38d9-4808-c77a-95d4-4963fe757ad6, 'name': SearchDatastore_Task, 'duration_secs': 0.011865} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.107344] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d16bfc23-13d6-4ee0-b5aa-237096055c0f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.113232] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 954.113232] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521c7732-a068-5270-0873-47679db09ea6" [ 954.113232] env[61936]: _type = "Task" [ 954.113232] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.117767] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Applying migration context for instance 9bd97f84-7c21-44ad-9ed0-d7e4097e264e as it has an incoming, in-progress migration 43339264-3f33-4305-8859-baeb09cac628. Migration status is post-migrating {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 954.119477] env[61936]: INFO nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating resource usage from migration 43339264-3f33-4305-8859-baeb09cac628 [ 954.127472] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521c7732-a068-5270-0873-47679db09ea6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.146404] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.146575] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 5526cbd5-b1ad-453b-8401-eee7aa356606 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.146731] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f2844173-4114-4158-99ad-ac651efbb607 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.146841] env[61936]: WARNING nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 10669072-a8e8-4262-b390-8bf31af6fafe is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 954.146958] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147090] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fa845f8e-957c-4c0b-a647-190c32989dcd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147212] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f3f2a0a8-cfe3-462f-a940-05580d5da32d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147477] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance b780470b-57d0-43e5-be52-539c78cada6b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147477] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 1d25e242-3542-4707-9112-3711e17df577 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147588] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Migration 43339264-3f33-4305-8859-baeb09cac628 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 954.147643] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 9bd97f84-7c21-44ad-9ed0-d7e4097e264e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147751] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 45fd1d4d-7125-43a7-afe7-53f737fb9f8e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147860] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance a99a6fef-3cef-409a-b001-4aca97f852c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.147972] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fbde880a-b47c-49e8-b84d-0efee21d62c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 954.151869] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-fa845f8e-957c-4c0b-a647-190c32989dcd" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.152609] env[61936]: DEBUG nova.objects.instance [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lazy-loading 'flavor' on Instance uuid fa845f8e-957c-4c0b-a647-190c32989dcd {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.165124] env[61936]: DEBUG nova.network.neutron [-] [instance: 1d25e242-3542-4707-9112-3711e17df577] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.252040] env[61936]: DEBUG oslo_vmware.api [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253204, 'name': PowerOffVM_Task, 'duration_secs': 0.267121} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.252040] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253205, 'name': ReconfigVM_Task, 'duration_secs': 0.228485} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.252040] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.252040] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.252040] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269950', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'name': 'volume-612c642f-abe9-4886-ac2f-baadfe0dcb78', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1a12fd59-bacd-42b3-99c1-8910f349e2cb', 'attached_at': '', 'detached_at': '', 'volume_id': '612c642f-abe9-4886-ac2f-baadfe0dcb78', 'serial': '612c642f-abe9-4886-ac2f-baadfe0dcb78'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 954.252682] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.252682] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd235703-476b-4238-87cd-6d322cbc85df {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.256226] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcba54e-4e5e-43ee-a2c9-9fb109891db2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.263515] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.263836] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0432aa7-54c2-4ef7-b59f-a21035b594e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.315310] env[61936]: DEBUG nova.compute.manager [req-844ee58e-e9b4-4f5e-b281-e224818cde68 req-03a4462b-a8e2-4449-a823-4b3847e398f1 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-vif-plugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 954.315682] env[61936]: DEBUG oslo_concurrency.lockutils [req-844ee58e-e9b4-4f5e-b281-e224818cde68 req-03a4462b-a8e2-4449-a823-4b3847e398f1 service nova] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.315771] env[61936]: DEBUG oslo_concurrency.lockutils [req-844ee58e-e9b4-4f5e-b281-e224818cde68 req-03a4462b-a8e2-4449-a823-4b3847e398f1 service nova] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.315929] env[61936]: DEBUG oslo_concurrency.lockutils [req-844ee58e-e9b4-4f5e-b281-e224818cde68 req-03a4462b-a8e2-4449-a823-4b3847e398f1 service nova] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.316099] env[61936]: DEBUG nova.compute.manager [req-844ee58e-e9b4-4f5e-b281-e224818cde68 req-03a4462b-a8e2-4449-a823-4b3847e398f1 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] No waiting events found dispatching network-vif-plugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 954.316593] env[61936]: WARNING nova.compute.manager [req-844ee58e-e9b4-4f5e-b281-e224818cde68 req-03a4462b-a8e2-4449-a823-4b3847e398f1 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received unexpected event network-vif-plugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 for instance with vm_state building and task_state spawning. [ 954.355617] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.356053] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.356280] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Deleting the datastore file [datastore1] f2844173-4114-4158-99ad-ac651efbb607 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.357472] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-647699cb-d24c-4db9-9476-26c50d1e8958 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.360010] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.360222] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.360396] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Deleting the datastore file [datastore1] 1a12fd59-bacd-42b3-99c1-8910f349e2cb {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.360640] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09de9bc0-e1e2-47de-967b-fe4f8ee726de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.369551] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for the task: (returnval){ [ 954.369551] env[61936]: value = "task-1253209" [ 954.369551] env[61936]: _type = "Task" [ 954.369551] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.369551] env[61936]: DEBUG oslo_vmware.api [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for the task: (returnval){ [ 954.369551] env[61936]: value = "task-1253208" [ 954.369551] env[61936]: _type = "Task" [ 954.369551] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.383644] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253209, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.387471] env[61936]: DEBUG oslo_vmware.api [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253208, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.437875] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253203, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649651} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.438126] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 45fd1d4d-7125-43a7-afe7-53f737fb9f8e/45fd1d4d-7125-43a7-afe7-53f737fb9f8e.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 954.438347] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.438657] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0eb37ab7-db52-480e-b036-2ee40f57811c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.446698] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 954.446698] env[61936]: value = "task-1253210" [ 954.446698] env[61936]: _type = "Task" [ 954.446698] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.456653] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253210, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.515494] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aad4321c-fe30-49a1-99a1-ba3136fa6f51 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.527898] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17cc8343-3643-446a-ab89-eab6135449b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.564482] env[61936]: DEBUG nova.compute.manager [req-7db5ee1c-f9cf-4fd2-99f4-f69bfa83f6d3 req-f6a5a8de-4463-4e12-bc14-0e9fbcd734b9 service nova] [instance: 1d25e242-3542-4707-9112-3711e17df577] Detach interface failed, port_id=bdcc6520-e3d6-482f-a1ee-2e9fbc9aed19, reason: Instance 1d25e242-3542-4707-9112-3711e17df577 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 954.624580] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521c7732-a068-5270-0873-47679db09ea6, 'name': SearchDatastore_Task, 'duration_secs': 0.024571} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.624856] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.625164] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] a99a6fef-3cef-409a-b001-4aca97f852c0/a99a6fef-3cef-409a-b001-4aca97f852c0.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.625437] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc3cb8bf-7322-4ad9-99b4-16213c9801b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.632487] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 954.632487] env[61936]: value = "task-1253211" [ 954.632487] env[61936]: _type = "Task" [ 954.632487] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.640694] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.652529] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 954.659221] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d360c449-49b0-4f2f-a7a3-791b602d7e04 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.680735] env[61936]: INFO nova.compute.manager [-] [instance: 1d25e242-3542-4707-9112-3711e17df577] Took 1.31 seconds to deallocate network for instance. [ 954.680811] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.682573] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51654fb2-1747-40c9-b4af-8bf73d2352ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.692019] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 954.692019] env[61936]: value = "task-1253212" [ 954.692019] env[61936]: _type = "Task" [ 954.692019] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.700605] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.825106] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.825507] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84d06825-b4b9-4e23-9c21-867faa7073fa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.838804] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 954.838804] env[61936]: value = "task-1253213" [ 954.838804] env[61936]: _type = "Task" [ 954.838804] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.848448] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253213, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.882691] env[61936]: DEBUG oslo_vmware.api [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Task: {'id': task-1253209, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169004} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.884069] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.884069] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.884069] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.889785] env[61936]: DEBUG oslo_vmware.api [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Task: {'id': task-1253208, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343005} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.889989] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 954.890106] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 954.890269] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 954.890452] env[61936]: INFO nova.compute.manager [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] [instance: f2844173-4114-4158-99ad-ac651efbb607] Took 1.20 seconds to destroy the instance on the hypervisor. [ 954.890880] env[61936]: DEBUG oslo.service.loopingcall [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.891031] env[61936]: DEBUG nova.compute.manager [-] [instance: f2844173-4114-4158-99ad-ac651efbb607] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 954.891129] env[61936]: DEBUG nova.network.neutron [-] [instance: f2844173-4114-4158-99ad-ac651efbb607] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.957382] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253210, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066411} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.957382] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.958535] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 954.959349] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f662f4-f3a9-4d58-be1c-a5a31c7a6666 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.962110] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2efc200a-2019-4a76-8df1-b990f701ea88 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.990217] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 45fd1d4d-7125-43a7-afe7-53f737fb9f8e/45fd1d4d-7125-43a7-afe7-53f737fb9f8e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.993263] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa471d38-b2da-41aa-8f3d-a7fb2c6905b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.004537] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-830ef9d5-37ce-4437-a640-57421b86cd0d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.031362] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 955.031362] env[61936]: value = "task-1253214" [ 955.031362] env[61936]: _type = "Task" [ 955.031362] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.050292] env[61936]: ERROR nova.compute.manager [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Failed to detach volume 612c642f-abe9-4886-ac2f-baadfe0dcb78 from /dev/sda: nova.exception.InstanceNotFound: Instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb could not be found. [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Traceback (most recent call last): [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self.driver.rebuild(**kwargs) [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] raise NotImplementedError() [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] NotImplementedError [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] During handling of the above exception, another exception occurred: [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Traceback (most recent call last): [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self.driver.detach_volume(context, old_connection_info, [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] return self._volumeops.detach_volume(connection_info, instance) [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self._detach_volume_vmdk(connection_info, instance) [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] stable_ref.fetch_moref(session) [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] raise exception.InstanceNotFound(instance_id=self._uuid) [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] nova.exception.InstanceNotFound: Instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb could not be found. [ 955.050292] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.058466] env[61936]: DEBUG nova.network.neutron [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Successfully updated port: d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.060024] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253214, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.144148] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253211, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.155825] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 955.178732] env[61936]: DEBUG nova.compute.manager [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 955.178966] env[61936]: DEBUG nova.compute.manager [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing instance network info cache due to event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 955.179891] env[61936]: DEBUG oslo_concurrency.lockutils [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] Acquiring lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.179891] env[61936]: DEBUG oslo_concurrency.lockutils [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] Acquired lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.179891] env[61936]: DEBUG nova.network.neutron [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing network info cache for port d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.189173] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.206754] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253212, 'name': PowerOffVM_Task, 'duration_secs': 0.266275} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.207644] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.213453] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Reconfiguring VM instance instance-0000004f to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 955.214380] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c0e9e0a-8566-4351-b6b2-55afe8ab5ed9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.238268] env[61936]: DEBUG nova.compute.utils [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Build of instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb aborted: Failed to rebuild volume backed instance. {{(pid=61936) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 955.242288] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 955.242288] env[61936]: value = "task-1253215" [ 955.242288] env[61936]: _type = "Task" [ 955.242288] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.243497] env[61936]: ERROR nova.compute.manager [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb aborted: Failed to rebuild volume backed instance. [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Traceback (most recent call last): [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self.driver.rebuild(**kwargs) [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] raise NotImplementedError() [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] NotImplementedError [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] During handling of the above exception, another exception occurred: [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Traceback (most recent call last): [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3642, in _rebuild_volume_backed_instance [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self._detach_root_volume(context, instance, root_bdm) [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3621, in _detach_root_volume [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] with excutils.save_and_reraise_exception(): [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self.force_reraise() [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] raise self.value [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self.driver.detach_volume(context, old_connection_info, [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] return self._volumeops.detach_volume(connection_info, instance) [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self._detach_volume_vmdk(connection_info, instance) [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] stable_ref.fetch_moref(session) [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] raise exception.InstanceNotFound(instance_id=self._uuid) [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] nova.exception.InstanceNotFound: Instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb could not be found. [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] During handling of the above exception, another exception occurred: [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Traceback (most recent call last): [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 11263, in _error_out_instance_on_exception [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] yield [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3910, in rebuild_instance [ 955.243497] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self._do_rebuild_instance_with_claim( [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3996, in _do_rebuild_instance_with_claim [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self._do_rebuild_instance( [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 4188, in _do_rebuild_instance [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self._rebuild_default_impl(**kwargs) [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3765, in _rebuild_default_impl [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] self._rebuild_volume_backed_instance( [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] File "/opt/stack/nova/nova/compute/manager.py", line 3657, in _rebuild_volume_backed_instance [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] raise exception.BuildAbortException( [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] nova.exception.BuildAbortException: Build of instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb aborted: Failed to rebuild volume backed instance. [ 955.244685] env[61936]: ERROR nova.compute.manager [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] [ 955.259490] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253215, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.339822] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.340121] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.340320] env[61936]: DEBUG nova.compute.manager [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Going to confirm migration 1 {{(pid=61936) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5174}} [ 955.351683] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253213, 'name': PowerOffVM_Task, 'duration_secs': 0.307673} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.351970] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.352792] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c39788f-8885-4fe7-8e92-2636521e0239 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.372905] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6c3579-f46c-4d09-8698-2b051660299f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.544292] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.563823] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.644332] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253211, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.739531} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.644646] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] a99a6fef-3cef-409a-b001-4aca97f852c0/a99a6fef-3cef-409a-b001-4aca97f852c0.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.644898] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.645203] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-157c5d44-b9ad-418e-87e5-c3d4f74ec2f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.652119] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 955.652119] env[61936]: value = "task-1253216" [ 955.652119] env[61936]: _type = "Task" [ 955.652119] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.661755] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance dc6523b7-c15c-469f-83a5-5c5dc5f2d230 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 955.663259] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253216, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.723509] env[61936]: DEBUG nova.network.neutron [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 955.759218] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253215, 'name': ReconfigVM_Task, 'duration_secs': 0.323621} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.759520] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Reconfigured VM instance instance-0000004f to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 955.759728] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 955.760320] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c41eba4-e496-4d44-8b07-c824ef420a5e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.767259] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 955.767259] env[61936]: value = "task-1253217" [ 955.767259] env[61936]: _type = "Task" [ 955.767259] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.779480] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253217, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.784694] env[61936]: DEBUG nova.network.neutron [-] [instance: f2844173-4114-4158-99ad-ac651efbb607] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.850845] env[61936]: DEBUG nova.network.neutron [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.885775] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Creating Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 955.886275] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-88c67c48-df4c-4871-bfdd-fd299caba288 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.897223] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 955.897223] env[61936]: value = "task-1253218" [ 955.897223] env[61936]: _type = "Task" [ 955.897223] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.909598] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253218, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.945356] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.945612] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.945918] env[61936]: DEBUG nova.network.neutron [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.946150] env[61936]: DEBUG nova.objects.instance [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lazy-loading 'info_cache' on Instance uuid 9bd97f84-7c21-44ad-9ed0-d7e4097e264e {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.983915] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525edfbc-d030-3015-a7b7-3bfc41a9a025/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 955.984841] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfd4198-4032-43ca-8fa6-4735eca76c3c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.992714] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525edfbc-d030-3015-a7b7-3bfc41a9a025/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 955.992901] env[61936]: ERROR oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525edfbc-d030-3015-a7b7-3bfc41a9a025/disk-0.vmdk due to incomplete transfer. [ 955.993157] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-899b5d65-0744-489b-b79e-bfc71ff1953a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.999714] env[61936]: DEBUG oslo_vmware.rw_handles [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525edfbc-d030-3015-a7b7-3bfc41a9a025/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 955.999914] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Uploaded image af56f59f-c70a-4b04-91e5-d2e24b68120b to the Glance image server {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 956.002178] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Destroying the VM {{(pid=61936) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 956.002439] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-32f33b10-9be9-47f5-84f9-6951d57f2125 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.008470] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 956.008470] env[61936]: value = "task-1253219" [ 956.008470] env[61936]: _type = "Task" [ 956.008470] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.017242] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253219, 'name': Destroy_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.035922] env[61936]: DEBUG nova.compute.manager [req-250d173a-90df-4844-bc4b-30977e48ec8f req-0c37fe98-985b-4bf3-9bbb-62d671fc144f service nova] [instance: f2844173-4114-4158-99ad-ac651efbb607] Received event network-vif-deleted-dd14ed80-5888-44af-897b-0ed4dd854e0d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 956.045958] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253214, 'name': ReconfigVM_Task, 'duration_secs': 0.596624} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.046294] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 45fd1d4d-7125-43a7-afe7-53f737fb9f8e/45fd1d4d-7125-43a7-afe7-53f737fb9f8e.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.047067] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0069daad-7b02-48ba-b1c4-ca81fa57f057 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.053024] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 956.053024] env[61936]: value = "task-1253220" [ 956.053024] env[61936]: _type = "Task" [ 956.053024] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.062618] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253220, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.163351] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253216, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070445} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.165778] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 0137305a-dace-4eda-9d90-7233c001176f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 956.165778] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 956.165778] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 956.169113] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.169113] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7565725-e839-4cac-9d34-bcb3a2457392 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.192916] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] a99a6fef-3cef-409a-b001-4aca97f852c0/a99a6fef-3cef-409a-b001-4aca97f852c0.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.196297] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16e73475-a7a9-4fcd-803d-4e14d4896ede {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.217953] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 956.217953] env[61936]: value = "task-1253221" [ 956.217953] env[61936]: _type = "Task" [ 956.217953] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.226721] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253221, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.279257] env[61936]: DEBUG oslo_vmware.api [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253217, 'name': PowerOnVM_Task, 'duration_secs': 0.428224} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.280743] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.280743] env[61936]: DEBUG nova.compute.manager [None req-c143ceef-8b4a-47ad-b734-366224209b0c tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 956.281126] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1f95c4-7d23-462d-bfd2-1b6d1584ecb9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.287287] env[61936]: INFO nova.compute.manager [-] [instance: f2844173-4114-4158-99ad-ac651efbb607] Took 1.40 seconds to deallocate network for instance. [ 956.355367] env[61936]: DEBUG oslo_concurrency.lockutils [req-e32f75ae-3e0b-462e-b186-6c5eb7ac377c req-d45d5232-89b1-46dd-8bf2-97ecc03c605b service nova] Releasing lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.355743] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.355902] env[61936]: DEBUG nova.network.neutron [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.408501] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253218, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.444317] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9382ae22-613b-4a23-8b23-a4214cf229aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.454345] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637e1654-c617-470b-a62c-a9e4e93d0e4b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.486126] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27c90e6-02d6-450e-bcad-a8eb211215a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.494431] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82eeeb98-bec1-43e2-8869-9f42a2876624 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.510014] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.518907] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253219, 'name': Destroy_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.564154] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253220, 'name': Rename_Task, 'duration_secs': 0.508478} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.564545] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.564862] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22517006-d15c-424b-a9d8-8d998c654d5e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.572122] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 956.572122] env[61936]: value = "task-1253222" [ 956.572122] env[61936]: _type = "Task" [ 956.572122] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.581738] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253222, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.728600] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253221, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.799848] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.889812] env[61936]: DEBUG nova.network.neutron [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.910552] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253218, 'name': CreateSnapshot_Task, 'duration_secs': 0.978933} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.910552] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Created Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 956.912030] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f05f8c-aa36-46e5-bd0f-c3d4f2664ee7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.015484] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 957.024305] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253219, 'name': Destroy_Task, 'duration_secs': 0.832407} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.026051] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Destroyed the VM [ 957.026051] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Deleting Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 957.026051] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-db75086e-264f-4a07-aafc-175720eadc65 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.031979] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 957.031979] env[61936]: value = "task-1253223" [ 957.031979] env[61936]: _type = "Task" [ 957.031979] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.036744] env[61936]: DEBUG nova.network.neutron [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [{"id": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "address": "fa:16:3e:65:44:c9", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd169ca50-5a", "ovs_interfaceid": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.043940] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253223, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.083030] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253222, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.228810] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253221, 'name': ReconfigVM_Task, 'duration_secs': 0.867761} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.229260] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Reconfigured VM instance instance-00000054 to attach disk [datastore2] a99a6fef-3cef-409a-b001-4aca97f852c0/a99a6fef-3cef-409a-b001-4aca97f852c0.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.229957] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1debe9db-956b-4b63-a391-3ad0fd917c74 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.236495] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 957.236495] env[61936]: value = "task-1253224" [ 957.236495] env[61936]: _type = "Task" [ 957.236495] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.245831] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253224, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.263412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.352702] env[61936]: DEBUG nova.network.neutron [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance_info_cache with network_info: [{"id": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "address": "fa:16:3e:34:7d:9b", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap432e9691-4e", "ovs_interfaceid": "432e9691-4e6f-4550-a7f4-7396d05e96d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.428598] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Creating linked-clone VM from snapshot {{(pid=61936) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 957.429531] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-57141514-6ccd-456b-9fb7-679a2369d11b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.437338] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 957.437338] env[61936]: value = "task-1253225" [ 957.437338] env[61936]: _type = "Task" [ 957.437338] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.445294] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253225, 'name': CloneVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.521091] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 957.521356] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.421s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.521668] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.776s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.523535] env[61936]: INFO nova.compute.claims [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.542177] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.542859] env[61936]: DEBUG nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance network_info: |[{"id": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "address": "fa:16:3e:65:44:c9", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd169ca50-5a", "ovs_interfaceid": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 957.546822] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:44:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd169ca50-5a14-4a65-9da0-ea79c80f75c8', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.556530] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating folder: Project (905821a1ff2b4011994c0d2d8bc08b13). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 957.557523] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "62b33dd9-e523-4309-94e6-59671fcfbaea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.557795] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.562098] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03870622-2b6b-43e4-a31a-46f8d5a02f94 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.567828] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253223, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.580078] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created folder: Project (905821a1ff2b4011994c0d2d8bc08b13) in parent group-v269874. [ 957.580078] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating folder: Instances. Parent ref: group-v269993. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 957.581128] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d6037a0c-5b4b-462c-a269-1e976bd04d81 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.586174] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253222, 'name': PowerOnVM_Task, 'duration_secs': 0.691434} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.586856] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.587087] env[61936]: INFO nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Took 11.18 seconds to spawn the instance on the hypervisor. [ 957.587279] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 957.588326] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01412e5-6927-4e20-ba90-ad400002f2fa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.595894] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created folder: Instances in parent group-v269993. [ 957.596171] env[61936]: DEBUG oslo.service.loopingcall [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.600783] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.601231] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0e097c1-a1d7-4412-bda4-c13664d643b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.620992] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.620992] env[61936]: value = "task-1253228" [ 957.620992] env[61936]: _type = "Task" [ 957.620992] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.629697] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253228, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.746948] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253224, 'name': Rename_Task, 'duration_secs': 0.189929} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.747282] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.747548] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1928c6de-af82-4737-b3e1-75f0bca6277e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.755048] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 957.755048] env[61936]: value = "task-1253229" [ 957.755048] env[61936]: _type = "Task" [ 957.755048] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.763463] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.859631] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-9bd97f84-7c21-44ad-9ed0-d7e4097e264e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.860048] env[61936]: DEBUG nova.objects.instance [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lazy-loading 'migration_context' on Instance uuid 9bd97f84-7c21-44ad-9ed0-d7e4097e264e {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.948074] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253225, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.043451] env[61936]: DEBUG oslo_vmware.api [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253223, 'name': RemoveSnapshot_Task, 'duration_secs': 0.910222} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.043801] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Deleted Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 958.044069] env[61936]: INFO nova.compute.manager [None req-669315a5-66d4-4147-a850-3faf27b4e01e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Took 15.76 seconds to snapshot the instance on the hypervisor. [ 958.067619] env[61936]: DEBUG nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 958.110065] env[61936]: INFO nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Took 37.70 seconds to build instance. [ 958.132860] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253228, 'name': CreateVM_Task, 'duration_secs': 0.337431} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.133059] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.133731] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.133901] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.134237] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.134487] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7237080d-bae3-4043-8f5c-ac7f73941aa9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.139173] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 958.139173] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5289bb26-a535-2720-d2ee-4f9c2a502b45" [ 958.139173] env[61936]: _type = "Task" [ 958.139173] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.149127] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5289bb26-a535-2720-d2ee-4f9c2a502b45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.265720] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253229, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.274326] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquiring lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.274564] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.274775] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquiring lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.274965] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.275204] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.277428] env[61936]: INFO nova.compute.manager [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Terminating instance [ 958.363065] env[61936]: DEBUG nova.objects.base [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Object Instance<9bd97f84-7c21-44ad-9ed0-d7e4097e264e> lazy-loaded attributes: info_cache,migration_context {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 958.363865] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71eb563b-4eb6-472a-8a2e-ab58569b5d87 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.385396] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aed431f5-65b9-421b-9b61-fcf555acc9d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.391047] env[61936]: DEBUG oslo_vmware.api [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 958.391047] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52646d28-f95d-7b89-581e-115ef6287fc4" [ 958.391047] env[61936]: _type = "Task" [ 958.391047] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.399690] env[61936]: DEBUG oslo_vmware.api [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52646d28-f95d-7b89-581e-115ef6287fc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.447214] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253225, 'name': CloneVM_Task} progress is 95%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.594023] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.611687] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.216s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.649665] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5289bb26-a535-2720-d2ee-4f9c2a502b45, 'name': SearchDatastore_Task, 'duration_secs': 0.033566} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.652551] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.652798] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.653044] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.653200] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.653379] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.653843] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b37cb55-237b-4227-8429-202d7b02d462 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.664151] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.664400] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.665152] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89966087-6f7c-4ebd-a1f9-fc6640b46c26 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.672496] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 958.672496] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5247062a-0b82-5c5d-ffba-c507fc69ab1a" [ 958.672496] env[61936]: _type = "Task" [ 958.672496] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.680393] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5247062a-0b82-5c5d-ffba-c507fc69ab1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.765925] env[61936]: DEBUG oslo_vmware.api [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253229, 'name': PowerOnVM_Task, 'duration_secs': 0.539471} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.768297] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.768614] env[61936]: INFO nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Took 9.97 seconds to spawn the instance on the hypervisor. [ 958.768814] env[61936]: DEBUG nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 958.769821] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d09681-ce9f-4e8b-89a4-cc38c5bd0c21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.780670] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56bf00b-e3c5-47a3-a9f8-f430817c0bb0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.783491] env[61936]: DEBUG nova.compute.manager [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 958.783668] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42386f70-6317-4149-b560-4214458b6e64 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.791863] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2935bde3-ac29-4a2f-9c81-25bee270b3c8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.800026] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d253a86-1ed7-4538-992e-aca5261164be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.839681] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72b234d-2cb8-4155-b806-ca4cdada0722 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.849048] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7992a9ec-8243-4415-b857-9b213d29b5d7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.862368] env[61936]: WARNING nova.virt.vmwareapi.driver [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb could not be found. [ 958.862622] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.863423] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7beadda-674d-4c9a-ada5-c309a8547967 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.876046] env[61936]: DEBUG nova.compute.provider_tree [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.879364] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e92516-66e7-4299-b7ed-3682ceb101be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.903794] env[61936]: DEBUG oslo_vmware.api [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52646d28-f95d-7b89-581e-115ef6287fc4, 'name': SearchDatastore_Task, 'duration_secs': 0.010687} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.915027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.915705] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb could not be found. [ 958.915705] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.915705] env[61936]: INFO nova.compute.manager [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Took 0.13 seconds to destroy the instance on the hypervisor. [ 958.916563] env[61936]: DEBUG oslo.service.loopingcall [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.916563] env[61936]: DEBUG nova.compute.manager [-] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 958.916563] env[61936]: DEBUG nova.network.neutron [-] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 958.950183] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253225, 'name': CloneVM_Task, 'duration_secs': 1.251775} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.950475] env[61936]: INFO nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Created linked-clone VM from snapshot [ 958.951255] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7dcc44-44df-420b-9183-cf4ae53833bc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.959153] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Uploading image 2f04c09d-c433-442e-b891-3fa38ecc215e {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 958.985940] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 958.985940] env[61936]: value = "vm-269992" [ 958.985940] env[61936]: _type = "VirtualMachine" [ 958.985940] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 958.986279] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9f7f0ae6-a3bb-4b5c-9864-5009e6e6e473 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.995532] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lease: (returnval){ [ 958.995532] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524db990-baee-676d-f4e6-810a0d94b092" [ 958.995532] env[61936]: _type = "HttpNfcLease" [ 958.995532] env[61936]: } obtained for exporting VM: (result){ [ 958.995532] env[61936]: value = "vm-269992" [ 958.995532] env[61936]: _type = "VirtualMachine" [ 958.995532] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 958.995866] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the lease: (returnval){ [ 958.995866] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524db990-baee-676d-f4e6-810a0d94b092" [ 958.995866] env[61936]: _type = "HttpNfcLease" [ 958.995866] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 959.006360] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.006360] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524db990-baee-676d-f4e6-810a0d94b092" [ 959.006360] env[61936]: _type = "HttpNfcLease" [ 959.006360] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.187705] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5247062a-0b82-5c5d-ffba-c507fc69ab1a, 'name': SearchDatastore_Task, 'duration_secs': 0.012833} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.188587] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71650c01-5d74-4871-a6fd-f0d7ace7bfcf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.195595] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 959.195595] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c89461-83c7-2a8e-20e0-88e406b8d25c" [ 959.195595] env[61936]: _type = "Task" [ 959.195595] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.204024] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c89461-83c7-2a8e-20e0-88e406b8d25c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.291532] env[61936]: INFO nova.compute.manager [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Took 38.83 seconds to build instance. [ 959.378804] env[61936]: DEBUG nova.scheduler.client.report [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.505557] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.505557] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524db990-baee-676d-f4e6-810a0d94b092" [ 959.505557] env[61936]: _type = "HttpNfcLease" [ 959.505557] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 959.505866] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 959.505866] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524db990-baee-676d-f4e6-810a0d94b092" [ 959.505866] env[61936]: _type = "HttpNfcLease" [ 959.505866] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 959.506682] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf90123-cc16-47fb-a5a4-7dff022d9c85 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.515851] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b1a17-1563-6e1f-7664-5f2d8a87ddf2/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 959.516065] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b1a17-1563-6e1f-7664-5f2d8a87ddf2/disk-0.vmdk for reading. {{(pid=61936) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 959.662611] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f2caf42a-3559-47ed-88c3-369631d63684 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.711248] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c89461-83c7-2a8e-20e0-88e406b8d25c, 'name': SearchDatastore_Task, 'duration_secs': 0.04304} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.711525] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.712432] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.712432] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2922497-886f-4e27-8a31-0b915a875c73 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.722320] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 959.722320] env[61936]: value = "task-1253231" [ 959.722320] env[61936]: _type = "Task" [ 959.722320] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.735187] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.793389] env[61936]: DEBUG oslo_concurrency.lockutils [None req-936bda7a-a823-4411-9d38-81c051b14dd6 tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.366s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.884486] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.885054] env[61936]: DEBUG nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 959.890695] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.406s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.892309] env[61936]: INFO nova.compute.claims [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.045026] env[61936]: DEBUG nova.compute.manager [req-4234a6cd-c28d-4b87-84e7-20af90ce703c req-f65a457b-612e-4757-abc5-5da5fb45eb83 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Received event network-vif-deleted-db50ae88-a45b-406e-855a-7124a33ce9ee {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 960.045124] env[61936]: INFO nova.compute.manager [req-4234a6cd-c28d-4b87-84e7-20af90ce703c req-f65a457b-612e-4757-abc5-5da5fb45eb83 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Neutron deleted interface db50ae88-a45b-406e-855a-7124a33ce9ee; detaching it from the instance and deleting it from the info cache [ 960.045283] env[61936]: DEBUG nova.network.neutron [req-4234a6cd-c28d-4b87-84e7-20af90ce703c req-f65a457b-612e-4757-abc5-5da5fb45eb83 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.102488] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.102662] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.102773] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.102969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.103186] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.105828] env[61936]: INFO nova.compute.manager [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Terminating instance [ 960.204966] env[61936]: DEBUG nova.network.neutron [-] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.237891] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253231, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.398747] env[61936]: DEBUG nova.compute.utils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.403874] env[61936]: DEBUG nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 960.404124] env[61936]: DEBUG nova.network.neutron [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 960.414278] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "a99a6fef-3cef-409a-b001-4aca97f852c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.414884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.415815] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "a99a6fef-3cef-409a-b001-4aca97f852c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.416826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.417282] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.426452] env[61936]: INFO nova.compute.manager [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Terminating instance [ 960.513025] env[61936]: DEBUG nova.policy [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cbdbed93b974583bae04b87b39e49a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ca963bd69594788806a9900942f895b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.554166] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b018efb0-f4e0-4bf0-977e-70101558a23e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.569456] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88f5d20-6483-426c-bc4e-f05278545d2b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.611210] env[61936]: DEBUG nova.compute.manager [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 960.611606] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.612666] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebad34de-37f8-4524-9230-a79d42c940fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.633748] env[61936]: DEBUG nova.compute.manager [req-4234a6cd-c28d-4b87-84e7-20af90ce703c req-f65a457b-612e-4757-abc5-5da5fb45eb83 service nova] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Detach interface failed, port_id=db50ae88-a45b-406e-855a-7124a33ce9ee, reason: Instance 1a12fd59-bacd-42b3-99c1-8910f349e2cb could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 960.647282] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.647900] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46729966-62e7-4812-b573-adddde68e897 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.656651] env[61936]: DEBUG oslo_vmware.api [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 960.656651] env[61936]: value = "task-1253232" [ 960.656651] env[61936]: _type = "Task" [ 960.656651] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.673998] env[61936]: DEBUG oslo_vmware.api [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253232, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.708580] env[61936]: INFO nova.compute.manager [-] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Took 1.79 seconds to deallocate network for instance. [ 960.738486] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253231, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.65054} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.738486] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.739098] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.739467] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae223bd1-5ffc-48eb-ac8e-df1df7c3549c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.749777] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 960.749777] env[61936]: value = "task-1253233" [ 960.749777] env[61936]: _type = "Task" [ 960.749777] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.762520] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253233, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.865756] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3556c161-abef-4842-9a9a-2487893ef3b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.875755] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44932a0f-adfa-41cb-903e-e2a55e6d773e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.916239] env[61936]: DEBUG nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 960.923466] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0afde43a-59cb-4f5a-88b4-b21b0a47dc2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.933325] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dba82b-de67-4062-8103-6a8fdda90d47 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.940466] env[61936]: DEBUG nova.compute.manager [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 960.940997] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.942126] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d3a36a-e501-4f69-a04e-f57b5b3ede2a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.960197] env[61936]: DEBUG nova.compute.provider_tree [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.965061] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.966983] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27e9cd80-2cb1-4ad0-bc86-9a9e0754bb98 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.975416] env[61936]: DEBUG oslo_vmware.api [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 960.975416] env[61936]: value = "task-1253234" [ 960.975416] env[61936]: _type = "Task" [ 960.975416] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.990336] env[61936]: DEBUG oslo_vmware.api [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.173123] env[61936]: DEBUG oslo_vmware.api [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253232, 'name': PowerOffVM_Task, 'duration_secs': 0.321363} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.173567] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.173815] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.174157] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebddea12-ab03-4b40-97d1-afa559bea1d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.258878] env[61936]: DEBUG nova.network.neutron [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Successfully created port: 746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.270034] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253233, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113574} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.270644] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.271201] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.271507] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.271709] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleting the datastore file [datastore2] 45fd1d4d-7125-43a7-afe7-53f737fb9f8e {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.272901] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb67576-d580-41ab-9963-22549538ab60 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.276589] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db571e3b-2ced-4046-b50d-466eb7664623 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.279091] env[61936]: INFO nova.compute.manager [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Took 0.57 seconds to detach 1 volumes for instance. [ 961.284013] env[61936]: DEBUG nova.compute.manager [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Deleting volume: 612c642f-abe9-4886-ac2f-baadfe0dcb78 {{(pid=61936) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 961.306606] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.308923] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-795abd03-18fe-477b-b694-c445e57d743d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.326762] env[61936]: DEBUG oslo_vmware.api [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 961.326762] env[61936]: value = "task-1253236" [ 961.326762] env[61936]: _type = "Task" [ 961.326762] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.334990] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 961.334990] env[61936]: value = "task-1253237" [ 961.334990] env[61936]: _type = "Task" [ 961.334990] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.343279] env[61936]: DEBUG oslo_vmware.api [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253236, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.357871] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253237, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.432254] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "484e5d41-29f4-4845-9633-157c03766978" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.432254] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.472267] env[61936]: DEBUG nova.scheduler.client.report [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 961.490917] env[61936]: DEBUG oslo_vmware.api [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253234, 'name': PowerOffVM_Task, 'duration_secs': 0.402641} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.491889] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.492325] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.493374] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abd8fe00-a106-4cc9-8031-95db0499e464 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.580321] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.580648] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.580925] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleting the datastore file [datastore2] a99a6fef-3cef-409a-b001-4aca97f852c0 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.581411] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12f535c2-c4aa-4f35-b6fc-19d08ab7d2ba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.593866] env[61936]: DEBUG oslo_vmware.api [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for the task: (returnval){ [ 961.593866] env[61936]: value = "task-1253240" [ 961.593866] env[61936]: _type = "Task" [ 961.593866] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.604441] env[61936]: DEBUG oslo_vmware.api [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253240, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.837665] env[61936]: DEBUG oslo_vmware.api [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253236, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19164} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.838104] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.838309] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.838672] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.838918] env[61936]: INFO nova.compute.manager [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Took 1.23 seconds to destroy the instance on the hypervisor. [ 961.839290] env[61936]: DEBUG oslo.service.loopingcall [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.839797] env[61936]: DEBUG nova.compute.manager [-] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 961.839797] env[61936]: DEBUG nova.network.neutron [-] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 961.856536] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.863805] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.928116] env[61936]: DEBUG nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 961.934110] env[61936]: DEBUG nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 961.961085] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 961.961251] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.961406] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 961.961582] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.961786] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 961.961926] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 961.962153] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 961.962309] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 961.962479] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 961.962640] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 961.964681] env[61936]: DEBUG nova.virt.hardware [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 961.964747] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c0d042-03e5-4237-b937-7ee52965549b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.974081] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.083s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.974852] env[61936]: DEBUG nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 961.984378] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.963s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.986060] env[61936]: INFO nova.compute.claims [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.995314] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1a620b-8f09-460c-b184-ee78b882f9bc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.106380] env[61936]: DEBUG oslo_vmware.api [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Task: {'id': task-1253240, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159395} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.107137] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.107255] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.107543] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.107778] env[61936]: INFO nova.compute.manager [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Took 1.17 seconds to destroy the instance on the hypervisor. [ 962.108064] env[61936]: DEBUG oslo.service.loopingcall [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.108342] env[61936]: DEBUG nova.compute.manager [-] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 962.108489] env[61936]: DEBUG nova.network.neutron [-] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.354279] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253237, 'name': ReconfigVM_Task, 'duration_secs': 0.606473} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.354585] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfigured VM instance instance-00000055 to attach disk [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.355264] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-169d1c86-ccaa-4956-b058-f33e418732aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.364028] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 962.364028] env[61936]: value = "task-1253241" [ 962.364028] env[61936]: _type = "Task" [ 962.364028] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.376253] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253241, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.401999] env[61936]: DEBUG nova.compute.manager [req-64762c06-f51e-4055-969a-17cad706a5b3 req-d90d8f07-9ab5-4428-8c24-70b8e1102ccc service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Received event network-vif-deleted-e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 962.402431] env[61936]: INFO nova.compute.manager [req-64762c06-f51e-4055-969a-17cad706a5b3 req-d90d8f07-9ab5-4428-8c24-70b8e1102ccc service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Neutron deleted interface e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53; detaching it from the instance and deleting it from the info cache [ 962.402682] env[61936]: DEBUG nova.network.neutron [req-64762c06-f51e-4055-969a-17cad706a5b3 req-d90d8f07-9ab5-4428-8c24-70b8e1102ccc service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.464911] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.491991] env[61936]: DEBUG nova.compute.utils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.500021] env[61936]: DEBUG nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 962.500021] env[61936]: DEBUG nova.network.neutron [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 962.603326] env[61936]: DEBUG nova.compute.manager [req-6d20386b-b177-4e73-9d93-13264356ced0 req-886ea9ce-c263-4b66-b636-a9ca03df8665 service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Received event network-vif-deleted-c29e5d63-c97e-4401-85ae-d7f0d13b72f0 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 962.603766] env[61936]: INFO nova.compute.manager [req-6d20386b-b177-4e73-9d93-13264356ced0 req-886ea9ce-c263-4b66-b636-a9ca03df8665 service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Neutron deleted interface c29e5d63-c97e-4401-85ae-d7f0d13b72f0; detaching it from the instance and deleting it from the info cache [ 962.603766] env[61936]: DEBUG nova.network.neutron [req-6d20386b-b177-4e73-9d93-13264356ced0 req-886ea9ce-c263-4b66-b636-a9ca03df8665 service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.631599] env[61936]: DEBUG nova.policy [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b05bdcace008487dba580405eb672ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90c29449db6c4beaa7f190d9225cb08a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 962.689011] env[61936]: DEBUG nova.network.neutron [-] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.875670] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253241, 'name': Rename_Task, 'duration_secs': 0.203228} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.876027] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.876306] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c1b212b-2e48-4d4c-95a4-724f8cc889c5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.883725] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 962.883725] env[61936]: value = "task-1253242" [ 962.883725] env[61936]: _type = "Task" [ 962.883725] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.891961] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.905789] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ba09230-43c2-4656-be57-66f43cd2e717 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.916025] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afbc38a-c942-4bb5-b839-8aa3e8676d74 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.945387] env[61936]: DEBUG nova.compute.manager [req-64762c06-f51e-4055-969a-17cad706a5b3 req-d90d8f07-9ab5-4428-8c24-70b8e1102ccc service nova] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Detach interface failed, port_id=e72c28ed-f0b9-4ae2-9f7f-4cf7dc6eea53, reason: Instance 45fd1d4d-7125-43a7-afe7-53f737fb9f8e could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 963.003315] env[61936]: DEBUG nova.network.neutron [-] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.004780] env[61936]: DEBUG nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 963.106179] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00233301-1709-4d38-acda-68d26fdaf704 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.124661] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a38461d-5902-42e5-874a-2cfeaa5c8c05 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.158995] env[61936]: DEBUG nova.compute.manager [req-6d20386b-b177-4e73-9d93-13264356ced0 req-886ea9ce-c263-4b66-b636-a9ca03df8665 service nova] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Detach interface failed, port_id=c29e5d63-c97e-4401-85ae-d7f0d13b72f0, reason: Instance a99a6fef-3cef-409a-b001-4aca97f852c0 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 963.191287] env[61936]: INFO nova.compute.manager [-] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Took 1.35 seconds to deallocate network for instance. [ 963.317537] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713ea7f0-71c9-4627-9f7b-57128e82c645 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.325972] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711b35bc-1c1c-46c1-8ef8-6790545bd757 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.358585] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ec30d8-4085-4b8c-8c5b-cd58a1dfae52 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.366385] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb8f9b0-7819-48b1-ac98-7796b9faf316 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.381688] env[61936]: DEBUG nova.compute.provider_tree [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.392839] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253242, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.408367] env[61936]: DEBUG nova.network.neutron [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Successfully created port: a3e152e9-3689-4f3d-ab48-c5f605ca1023 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.508095] env[61936]: INFO nova.compute.manager [-] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Took 1.40 seconds to deallocate network for instance. [ 963.701594] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.738646] env[61936]: DEBUG nova.network.neutron [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Successfully updated port: 746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.884153] env[61936]: DEBUG nova.scheduler.client.report [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 963.898110] env[61936]: DEBUG oslo_vmware.api [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253242, 'name': PowerOnVM_Task, 'duration_secs': 0.698586} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.898369] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.898592] env[61936]: INFO nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Took 10.69 seconds to spawn the instance on the hypervisor. [ 963.898784] env[61936]: DEBUG nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 963.899610] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80571b3f-1a6b-4d7d-9772-6a48b2c12024 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.016905] env[61936]: DEBUG nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 964.022599] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.051464] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 964.051783] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.051980] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 964.052224] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.052392] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 964.054035] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 964.054035] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 964.054035] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 964.054035] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 964.054035] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 964.054035] env[61936]: DEBUG nova.virt.hardware [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 964.054641] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edca87fb-4b33-4c35-aa0d-aa7fd3c27654 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.065671] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57538d4-73fa-4ed2-9288-785e44f8c782 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.241603] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.241757] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.241901] env[61936]: DEBUG nova.network.neutron [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.393922] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.394518] env[61936]: DEBUG nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 964.397708] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.386s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.399343] env[61936]: INFO nova.compute.claims [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.424190] env[61936]: INFO nova.compute.manager [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Took 41.60 seconds to build instance. [ 964.636311] env[61936]: DEBUG nova.compute.manager [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Received event network-vif-plugged-746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 964.636796] env[61936]: DEBUG oslo_concurrency.lockutils [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.636796] env[61936]: DEBUG oslo_concurrency.lockutils [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.636796] env[61936]: DEBUG oslo_concurrency.lockutils [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.636960] env[61936]: DEBUG nova.compute.manager [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] No waiting events found dispatching network-vif-plugged-746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 964.637105] env[61936]: WARNING nova.compute.manager [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Received unexpected event network-vif-plugged-746d118b-a889-44a8-a6fe-1f8e1c483c33 for instance with vm_state building and task_state spawning. [ 964.637274] env[61936]: DEBUG nova.compute.manager [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Received event network-changed-746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 964.637430] env[61936]: DEBUG nova.compute.manager [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Refreshing instance network info cache due to event network-changed-746d118b-a889-44a8-a6fe-1f8e1c483c33. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 964.637598] env[61936]: DEBUG oslo_concurrency.lockutils [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] Acquiring lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.800038] env[61936]: DEBUG nova.network.neutron [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.904178] env[61936]: DEBUG nova.compute.utils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.911729] env[61936]: DEBUG nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 964.911914] env[61936]: DEBUG nova.network.neutron [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 964.927292] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d2acb4d-abb4-4c20-b10c-25a244007bf2 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.114s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.963158] env[61936]: DEBUG nova.policy [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcee2be71ae044338df4422a3c575ca4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e81b119251964bdb8e5a0ea84b29f2a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.022815] env[61936]: DEBUG nova.network.neutron [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating instance_info_cache with network_info: [{"id": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "address": "fa:16:3e:ce:3d:0e", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746d118b-a8", "ovs_interfaceid": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.238785] env[61936]: DEBUG nova.network.neutron [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Successfully created port: 82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.412506] env[61936]: DEBUG nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 965.525441] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.525770] env[61936]: DEBUG nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Instance network_info: |[{"id": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "address": "fa:16:3e:ce:3d:0e", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746d118b-a8", "ovs_interfaceid": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 965.526337] env[61936]: DEBUG oslo_concurrency.lockutils [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] Acquired lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.527131] env[61936]: DEBUG nova.network.neutron [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Refreshing network info cache for port 746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.527988] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:3d:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd72ef32-a57c-43b0-93df-e8a030987d44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '746d118b-a889-44a8-a6fe-1f8e1c483c33', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.535800] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Creating folder: Project (9ca963bd69594788806a9900942f895b). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 965.536202] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e048f30d-eca9-4151-ba3c-13152c41b0a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.550901] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Created folder: Project (9ca963bd69594788806a9900942f895b) in parent group-v269874. [ 965.551928] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Creating folder: Instances. Parent ref: group-v269996. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 965.551928] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-414b3679-055d-4dc2-a544-23f64cbc638d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.561937] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Created folder: Instances in parent group-v269996. [ 965.562075] env[61936]: DEBUG oslo.service.loopingcall [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.562206] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.562975] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94c469c8-9770-45e5-a405-b8f7a3badf79 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.590038] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.590038] env[61936]: value = "task-1253245" [ 965.590038] env[61936]: _type = "Task" [ 965.590038] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.597344] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253245, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.718971] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bfea3d-445d-4b3b-8a13-3101ef624628 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.727444] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117ad3e4-cc5b-4d1c-b908-91de8d2a1652 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.764114] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbf4eb1-df40-45b9-b343-785c3ae03811 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.772857] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5366ea00-928d-4afd-a386-8eecaf839073 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.788438] env[61936]: DEBUG nova.compute.provider_tree [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.828725] env[61936]: DEBUG nova.network.neutron [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Successfully updated port: a3e152e9-3689-4f3d-ab48-c5f605ca1023 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.101567] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253245, 'name': CreateVM_Task, 'duration_secs': 0.389087} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.101813] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.102960] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.102960] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.103200] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.103488] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4894c858-aeee-4c4d-9caa-c0e5b2547e8e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.108301] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 966.108301] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525df2b5-194d-6af9-1d38-7e641fdf1468" [ 966.108301] env[61936]: _type = "Task" [ 966.108301] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.116595] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525df2b5-194d-6af9-1d38-7e641fdf1468, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.291679] env[61936]: DEBUG nova.scheduler.client.report [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 966.331049] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.331211] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.331365] env[61936]: DEBUG nova.network.neutron [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.394507] env[61936]: DEBUG nova.network.neutron [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updated VIF entry in instance network info cache for port 746d118b-a889-44a8-a6fe-1f8e1c483c33. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.394711] env[61936]: DEBUG nova.network.neutron [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating instance_info_cache with network_info: [{"id": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "address": "fa:16:3e:ce:3d:0e", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746d118b-a8", "ovs_interfaceid": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.427030] env[61936]: DEBUG nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 966.454319] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 966.454506] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.454666] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 966.454857] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.455018] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 966.455182] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 966.455530] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 966.455617] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 966.455706] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 966.455869] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 966.456078] env[61936]: DEBUG nova.virt.hardware [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 966.456972] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0019435f-1f25-406e-9be3-71bde4271fa8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.465286] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f87d914-66db-483b-8897-9144f7f89389 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.620278] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525df2b5-194d-6af9-1d38-7e641fdf1468, 'name': SearchDatastore_Task, 'duration_secs': 0.010294} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.620619] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.620939] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.621208] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.621360] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.621601] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.621813] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cbb87e9c-988a-449b-a7b2-a4a855e40c9a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.630525] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.630727] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.631519] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c55637d-794f-4bfc-961b-d6cd5c497753 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.637035] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 966.637035] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528251d8-2848-adae-b33d-045530f89cc5" [ 966.637035] env[61936]: _type = "Task" [ 966.637035] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.645704] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528251d8-2848-adae-b33d-045530f89cc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.670019] env[61936]: DEBUG nova.compute.manager [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Received event network-vif-plugged-a3e152e9-3689-4f3d-ab48-c5f605ca1023 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 966.671228] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Acquiring lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.671228] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.671228] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.671228] env[61936]: DEBUG nova.compute.manager [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] No waiting events found dispatching network-vif-plugged-a3e152e9-3689-4f3d-ab48-c5f605ca1023 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 966.671228] env[61936]: WARNING nova.compute.manager [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Received unexpected event network-vif-plugged-a3e152e9-3689-4f3d-ab48-c5f605ca1023 for instance with vm_state building and task_state spawning. [ 966.671491] env[61936]: DEBUG nova.compute.manager [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Received event network-changed-a3e152e9-3689-4f3d-ab48-c5f605ca1023 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 966.671491] env[61936]: DEBUG nova.compute.manager [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Refreshing instance network info cache due to event network-changed-a3e152e9-3689-4f3d-ab48-c5f605ca1023. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 966.671714] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Acquiring lock "refresh_cache-8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.681695] env[61936]: DEBUG nova.compute.manager [req-7d0b0d3c-de1b-449a-8c34-17a1fb072d6d req-bccd4f24-8134-4062-a9a6-c03f159bf5c7 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Received event network-vif-plugged-82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 966.681964] env[61936]: DEBUG oslo_concurrency.lockutils [req-7d0b0d3c-de1b-449a-8c34-17a1fb072d6d req-bccd4f24-8134-4062-a9a6-c03f159bf5c7 service nova] Acquiring lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.682560] env[61936]: DEBUG oslo_concurrency.lockutils [req-7d0b0d3c-de1b-449a-8c34-17a1fb072d6d req-bccd4f24-8134-4062-a9a6-c03f159bf5c7 service nova] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.682767] env[61936]: DEBUG oslo_concurrency.lockutils [req-7d0b0d3c-de1b-449a-8c34-17a1fb072d6d req-bccd4f24-8134-4062-a9a6-c03f159bf5c7 service nova] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.682975] env[61936]: DEBUG nova.compute.manager [req-7d0b0d3c-de1b-449a-8c34-17a1fb072d6d req-bccd4f24-8134-4062-a9a6-c03f159bf5c7 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] No waiting events found dispatching network-vif-plugged-82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 966.683388] env[61936]: WARNING nova.compute.manager [req-7d0b0d3c-de1b-449a-8c34-17a1fb072d6d req-bccd4f24-8134-4062-a9a6-c03f159bf5c7 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Received unexpected event network-vif-plugged-82ce2f0e-a794-4a8a-a494-0b9deb336108 for instance with vm_state building and task_state spawning. [ 966.786755] env[61936]: DEBUG nova.network.neutron [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Successfully updated port: 82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.797977] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.797977] env[61936]: DEBUG nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 966.800895] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.701s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.801209] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.803315] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.614s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.803534] env[61936]: DEBUG nova.objects.instance [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lazy-loading 'resources' on Instance uuid 1d25e242-3542-4707-9112-3711e17df577 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.826953] env[61936]: INFO nova.scheduler.client.report [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted allocations for instance 10669072-a8e8-4262-b390-8bf31af6fafe [ 966.864686] env[61936]: DEBUG nova.network.neutron [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 966.897872] env[61936]: DEBUG oslo_concurrency.lockutils [req-5cd4bd26-5e2a-409c-9661-bc64b14ccca2 req-e1ba6d30-309d-48b1-8693-ecfba79c3359 service nova] Releasing lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.999713] env[61936]: DEBUG nova.network.neutron [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Updating instance_info_cache with network_info: [{"id": "a3e152e9-3689-4f3d-ab48-c5f605ca1023", "address": "fa:16:3e:45:59:ca", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3e152e9-36", "ovs_interfaceid": "a3e152e9-3689-4f3d-ab48-c5f605ca1023", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.148431] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528251d8-2848-adae-b33d-045530f89cc5, 'name': SearchDatastore_Task, 'duration_secs': 0.008669} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.149370] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be347d40-805d-41eb-8a98-251722fdd13c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.154931] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 967.154931] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520f394d-ec7a-3034-27b6-2384df3df32a" [ 967.154931] env[61936]: _type = "Task" [ 967.154931] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.162955] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520f394d-ec7a-3034-27b6-2384df3df32a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.291959] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.291959] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.291959] env[61936]: DEBUG nova.network.neutron [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.304944] env[61936]: DEBUG nova.compute.utils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.306533] env[61936]: DEBUG nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 967.306871] env[61936]: DEBUG nova.network.neutron [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 967.338144] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e027dc39-0bc0-4563-9671-c55b36f11189 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "10669072-a8e8-4262-b390-8bf31af6fafe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.711s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.363308] env[61936]: DEBUG nova.policy [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.502324] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.504788] env[61936]: DEBUG nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Instance network_info: |[{"id": "a3e152e9-3689-4f3d-ab48-c5f605ca1023", "address": "fa:16:3e:45:59:ca", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3e152e9-36", "ovs_interfaceid": "a3e152e9-3689-4f3d-ab48-c5f605ca1023", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 967.504788] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Acquired lock "refresh_cache-8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.504788] env[61936]: DEBUG nova.network.neutron [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Refreshing network info cache for port a3e152e9-3689-4f3d-ab48-c5f605ca1023 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.504788] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:59:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3e152e9-3689-4f3d-ab48-c5f605ca1023', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.515026] env[61936]: DEBUG oslo.service.loopingcall [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.521821] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.522549] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7ec3577-2196-4016-8cee-39988f216f62 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.560322] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.560322] env[61936]: value = "task-1253246" [ 967.560322] env[61936]: _type = "Task" [ 967.560322] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.571080] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253246, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.645157] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b1a17-1563-6e1f-7664-5f2d8a87ddf2/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 967.646598] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138c4afa-2daa-47f2-b110-99432d75768c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.653872] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b1a17-1563-6e1f-7664-5f2d8a87ddf2/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 967.654060] env[61936]: ERROR oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b1a17-1563-6e1f-7664-5f2d8a87ddf2/disk-0.vmdk due to incomplete transfer. [ 967.656729] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d1066cd8-c00b-452a-b179-6cd5c8c9e5ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.668483] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520f394d-ec7a-3034-27b6-2384df3df32a, 'name': SearchDatastore_Task, 'duration_secs': 0.013965} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.669719] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.670055] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8/0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.670388] env[61936]: DEBUG oslo_vmware.rw_handles [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524b1a17-1563-6e1f-7664-5f2d8a87ddf2/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 967.674020] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Uploaded image 2f04c09d-c433-442e-b891-3fa38ecc215e to the Glance image server {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 967.676042] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Destroying the VM {{(pid=61936) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 967.677584] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9bdc8c57-cc65-4d73-b160-d278f3f1cfe3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.679294] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7c3d675f-7df8-4c74-a006-8e73c7b65e55 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.681323] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8194632-4851-4122-b5cd-a37a262a4dd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.690729] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93400a8b-16d1-4576-99b7-7ab65c6074b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.694080] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 967.694080] env[61936]: value = "task-1253247" [ 967.694080] env[61936]: _type = "Task" [ 967.694080] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.694349] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 967.694349] env[61936]: value = "task-1253248" [ 967.694349] env[61936]: _type = "Task" [ 967.694349] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.727980] env[61936]: DEBUG nova.network.neutron [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Successfully created port: caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.731270] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2502ae18-babe-4ad1-8933-bb39d6337b31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.740277] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253247, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.740522] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253248, 'name': Destroy_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.745230] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cad50c-42ef-4ee6-81dd-3d071d839bec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.761982] env[61936]: DEBUG nova.compute.provider_tree [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.813274] env[61936]: DEBUG nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 967.944017] env[61936]: DEBUG nova.network.neutron [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.071845] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253246, 'name': CreateVM_Task, 'duration_secs': 0.397419} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.072028] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.072878] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.073108] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.073632] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.073959] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c577d47-9618-4652-a5b4-98a68bbe17c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.081489] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 968.081489] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b30dc9-b3c8-46f5-4975-a5cd5ff877bc" [ 968.081489] env[61936]: _type = "Task" [ 968.081489] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.094721] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b30dc9-b3c8-46f5-4975-a5cd5ff877bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.219422] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253247, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.219422] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253248, 'name': Destroy_Task, 'duration_secs': 0.390476} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.219422] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Destroyed the VM [ 968.219422] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Deleting Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 968.219422] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-07ab7ea9-fe28-490b-a2c9-61e72f99e536 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.225359] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 968.225359] env[61936]: value = "task-1253249" [ 968.225359] env[61936]: _type = "Task" [ 968.225359] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.234910] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253249, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.266835] env[61936]: DEBUG nova.scheduler.client.report [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 968.333464] env[61936]: DEBUG nova.network.neutron [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Updating instance_info_cache with network_info: [{"id": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "address": "fa:16:3e:ee:f9:9e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82ce2f0e-a7", "ovs_interfaceid": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.351565] env[61936]: DEBUG nova.network.neutron [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Updated VIF entry in instance network info cache for port a3e152e9-3689-4f3d-ab48-c5f605ca1023. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.351934] env[61936]: DEBUG nova.network.neutron [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Updating instance_info_cache with network_info: [{"id": "a3e152e9-3689-4f3d-ab48-c5f605ca1023", "address": "fa:16:3e:45:59:ca", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3e152e9-36", "ovs_interfaceid": "a3e152e9-3689-4f3d-ab48-c5f605ca1023", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.592830] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b30dc9-b3c8-46f5-4975-a5cd5ff877bc, 'name': SearchDatastore_Task, 'duration_secs': 0.05274} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.593162] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.593409] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.593755] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.593923] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.594121] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.594396] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c8866e8-85d9-460c-8e6a-de23499be4d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.604939] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.605448] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.607412] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59beb8a1-55d2-4697-a485-c8c99c8a15c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.613720] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 968.613720] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239ed44-17e9-6677-462e-717880b33aa1" [ 968.613720] env[61936]: _type = "Task" [ 968.613720] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.623508] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239ed44-17e9-6677-462e-717880b33aa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.706508] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253247, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544381} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.706825] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8/0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.707057] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.707500] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef05ef69-1b1c-41ec-8d36-1ad0af2d002b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.719020] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 968.719020] env[61936]: value = "task-1253250" [ 968.719020] env[61936]: _type = "Task" [ 968.719020] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.727602] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.736961] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253249, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.776179] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.973s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.778727] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.979s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.778972] env[61936]: DEBUG nova.objects.instance [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lazy-loading 'resources' on Instance uuid f2844173-4114-4158-99ad-ac651efbb607 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.807388] env[61936]: INFO nova.scheduler.client.report [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Deleted allocations for instance 1d25e242-3542-4707-9112-3711e17df577 [ 968.811272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.811404] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.825723] env[61936]: DEBUG nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 968.836439] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.836439] env[61936]: DEBUG nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Instance network_info: |[{"id": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "address": "fa:16:3e:ee:f9:9e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82ce2f0e-a7", "ovs_interfaceid": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 968.836439] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:f9:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82ce2f0e-a794-4a8a-a494-0b9deb336108', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.844759] env[61936]: DEBUG oslo.service.loopingcall [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.845258] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 968.845993] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a546da49-442b-4f97-b76a-1dba78dc3512 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.864194] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Releasing lock "refresh_cache-8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.864441] env[61936]: DEBUG nova.compute.manager [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 968.864610] env[61936]: DEBUG nova.compute.manager [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing instance network info cache due to event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 968.865012] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Acquiring lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.865192] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Acquired lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.865361] env[61936]: DEBUG nova.network.neutron [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing network info cache for port d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.867999] env[61936]: DEBUG nova.compute.manager [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Received event network-changed-82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 968.868198] env[61936]: DEBUG nova.compute.manager [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Refreshing instance network info cache due to event network-changed-82ce2f0e-a794-4a8a-a494-0b9deb336108. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 968.868395] env[61936]: DEBUG oslo_concurrency.lockutils [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] Acquiring lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.868563] env[61936]: DEBUG oslo_concurrency.lockutils [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] Acquired lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.868730] env[61936]: DEBUG nova.network.neutron [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Refreshing network info cache for port 82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.875905] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.875905] env[61936]: value = "task-1253251" [ 968.875905] env[61936]: _type = "Task" [ 968.875905] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.877948] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 968.878437] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.878437] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 968.878598] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.878687] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 968.878836] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 968.879075] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 968.879246] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 968.879412] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 968.879576] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 968.879746] env[61936]: DEBUG nova.virt.hardware [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 968.881162] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d12cb50-8d8e-42a2-9542-c7d90852d504 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.895311] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7d4f78-165e-443b-8d5b-0739b032bf94 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.899403] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253251, 'name': CreateVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.124377] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5239ed44-17e9-6677-462e-717880b33aa1, 'name': SearchDatastore_Task, 'duration_secs': 0.017507} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.125229] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b65a2d14-59fa-4c74-b2cd-303a15ffd66e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.130573] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 969.130573] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bee85c-a749-c519-47fe-01281ebcbee3" [ 969.130573] env[61936]: _type = "Task" [ 969.130573] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.139345] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bee85c-a749-c519-47fe-01281ebcbee3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.225920] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087323} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.225920] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.226496] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33eb0f8c-0627-4cf6-8f34-bf057942c01d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.251287] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8/0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.256317] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faf2d11b-5ca8-4271-b8e4-3e293187d748 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.270262] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253249, 'name': RemoveSnapshot_Task, 'duration_secs': 0.593658} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.270540] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Deleted Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 969.270902] env[61936]: DEBUG nova.compute.manager [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 969.272031] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805d1c27-0480-4f44-b3e6-bd9fea5a4d50 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.275961] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 969.275961] env[61936]: value = "task-1253252" [ 969.275961] env[61936]: _type = "Task" [ 969.275961] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.290037] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253252, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.315965] env[61936]: INFO nova.compute.manager [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Detaching volume 56048ceb-814a-4561-9098-d35c809208f0 [ 969.320579] env[61936]: DEBUG oslo_concurrency.lockutils [None req-562dab7e-7c63-4828-8da2-014a4beaa943 tempest-ServerMetadataTestJSON-112041471 tempest-ServerMetadataTestJSON-112041471-project-member] Lock "1d25e242-3542-4707-9112-3711e17df577" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.591s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.359669] env[61936]: INFO nova.virt.block_device [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Attempting to driver detach volume 56048ceb-814a-4561-9098-d35c809208f0 from mountpoint /dev/sdb [ 969.359781] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 969.359886] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269943', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'name': 'volume-56048ceb-814a-4561-9098-d35c809208f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377', 'attached_at': '', 'detached_at': '', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'serial': '56048ceb-814a-4561-9098-d35c809208f0'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 969.361537] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa166308-aa71-4bae-b64c-fb0399774007 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.391188] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb17a32-188a-44ec-843a-9a8b68a02efa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.401618] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c745da52-9964-431a-b041-5ba73574c7e5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.404040] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253251, 'name': CreateVM_Task, 'duration_secs': 0.360319} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.406305] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 969.407431] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.407598] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.408446] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.408446] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2df4e29a-7fbb-48d5-b2cd-c2dfe6b0f772 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.430663] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fac1796-6563-47f9-b01b-2c61e8b8c966 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.436355] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 969.436355] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd7c7c-a22d-de5b-35d6-4ec914183e86" [ 969.436355] env[61936]: _type = "Task" [ 969.436355] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.450460] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] The volume has not been displaced from its original location: [datastore2] volume-56048ceb-814a-4561-9098-d35c809208f0/volume-56048ceb-814a-4561-9098-d35c809208f0.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 969.455847] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Reconfiguring VM instance instance-00000035 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 969.458961] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5933668-8679-44ff-9245-546cca002fd9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.476765] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd7c7c-a22d-de5b-35d6-4ec914183e86, 'name': SearchDatastore_Task, 'duration_secs': 0.012954} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.477575] env[61936]: DEBUG nova.network.neutron [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Successfully updated port: caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.479306] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.479391] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 969.479584] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.483721] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 969.483721] env[61936]: value = "task-1253253" [ 969.483721] env[61936]: _type = "Task" [ 969.483721] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.491954] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253253, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.596109] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71060674-2e30-41df-aff6-4b35dce3d989 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.603806] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e77e7e-c325-4993-a05a-9ef8d085895b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.643941] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af29ce53-de25-4c43-abdf-db7a08cfe276 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.654629] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bee85c-a749-c519-47fe-01281ebcbee3, 'name': SearchDatastore_Task, 'duration_secs': 0.009082} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.655310] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.655433] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16/8d6c05f2-4bac-41ce-a9f0-4df3d86eba16.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.656628] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2613982a-f019-47f9-aec0-add3af9d2923 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.660544] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.661315] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.661315] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4567fa00-5a01-4168-afc0-50093005bdf0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.663417] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0463537-c1f1-49c2-bece-358dd12b3792 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.676043] env[61936]: DEBUG nova.compute.provider_tree [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.680432] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.680432] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 969.680432] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 969.680432] env[61936]: value = "task-1253254" [ 969.680432] env[61936]: _type = "Task" [ 969.680432] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.680432] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6b971e5-98df-4b35-8faa-af66a695dcb7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.688869] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 969.688869] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b95fe0-5011-645c-8594-7958c7cdc4d2" [ 969.688869] env[61936]: _type = "Task" [ 969.688869] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.691951] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253254, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.700909] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b95fe0-5011-645c-8594-7958c7cdc4d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.789624] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253252, 'name': ReconfigVM_Task, 'duration_secs': 0.323646} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.790091] env[61936]: INFO nova.compute.manager [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Shelve offloading [ 969.791353] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8/0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.792730] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e926738-a383-4dbb-bc66-07ca0b188f15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.798666] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 969.798666] env[61936]: value = "task-1253255" [ 969.798666] env[61936]: _type = "Task" [ 969.798666] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.810592] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253255, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.983223] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.983223] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.983662] env[61936]: DEBUG nova.network.neutron [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.996014] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253253, 'name': ReconfigVM_Task, 'duration_secs': 0.30017} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.997116] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Reconfigured VM instance instance-00000035 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 970.005015] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58bf3fc2-8654-48e7-b8e9-c5e002d91f19 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.017177] env[61936]: DEBUG nova.network.neutron [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Updated VIF entry in instance network info cache for port 82ce2f0e-a794-4a8a-a494-0b9deb336108. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.020027] env[61936]: DEBUG nova.network.neutron [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Updating instance_info_cache with network_info: [{"id": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "address": "fa:16:3e:ee:f9:9e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82ce2f0e-a7", "ovs_interfaceid": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.026153] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 970.026153] env[61936]: value = "task-1253256" [ 970.026153] env[61936]: _type = "Task" [ 970.026153] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.037928] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253256, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.093289] env[61936]: DEBUG nova.network.neutron [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updated VIF entry in instance network info cache for port d169ca50-5a14-4a65-9da0-ea79c80f75c8. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.093815] env[61936]: DEBUG nova.network.neutron [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [{"id": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "address": "fa:16:3e:65:44:c9", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd169ca50-5a", "ovs_interfaceid": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.182129] env[61936]: DEBUG nova.scheduler.client.report [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 970.194504] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253254, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498045} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.197515] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16/8d6c05f2-4bac-41ce-a9f0-4df3d86eba16.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.197778] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.198574] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02aa5ca1-fe1c-4b40-92b2-9b0f95fc7ac6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.206096] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b95fe0-5011-645c-8594-7958c7cdc4d2, 'name': SearchDatastore_Task, 'duration_secs': 0.020072} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.209984] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 970.209984] env[61936]: value = "task-1253257" [ 970.209984] env[61936]: _type = "Task" [ 970.209984] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.209984] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-726d50cc-2304-45dd-81f6-3df39aa28e8e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.217319] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 970.217319] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c87a64-a252-ba9d-ef05-f4e4c6892313" [ 970.217319] env[61936]: _type = "Task" [ 970.217319] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.220035] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253257, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.227881] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c87a64-a252-ba9d-ef05-f4e4c6892313, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.295786] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.296155] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f501f01a-a696-44dc-b898-638568b4d1c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.303860] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 970.303860] env[61936]: value = "task-1253258" [ 970.303860] env[61936]: _type = "Task" [ 970.303860] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.310622] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253255, 'name': Rename_Task, 'duration_secs': 0.185148} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.311376] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 970.311634] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-496516de-dedc-4744-9325-de4fd6d8a6af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.316432] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 970.316688] env[61936]: DEBUG nova.compute.manager [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 970.317478] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c23d82-95fe-46c3-b1d0-2886a9122db8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.320832] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 970.320832] env[61936]: value = "task-1253259" [ 970.320832] env[61936]: _type = "Task" [ 970.320832] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.329632] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.329632] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.329632] env[61936]: DEBUG nova.network.neutron [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.331587] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253259, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.517126] env[61936]: DEBUG nova.network.neutron [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 970.520946] env[61936]: DEBUG oslo_concurrency.lockutils [req-8b137a74-d529-4fe3-85c0-2b0d2bec45fc req-5d709c4f-4006-4fb6-95f9-591790a648c1 service nova] Releasing lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.538370] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253256, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.596892] env[61936]: DEBUG oslo_concurrency.lockutils [req-90898284-7e56-4400-9cba-c64ba52ed6de req-b02a7605-2088-4640-96d3-d508ea5514d2 service nova] Releasing lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.669602] env[61936]: DEBUG nova.network.neutron [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.690265] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.692619] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.429s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.719708] env[61936]: INFO nova.scheduler.client.report [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Deleted allocations for instance f2844173-4114-4158-99ad-ac651efbb607 [ 970.727203] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253257, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091776} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.730192] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.733906] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46e44d5-b7d2-439c-84c1-707d13b4afd5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.742035] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c87a64-a252-ba9d-ef05-f4e4c6892313, 'name': SearchDatastore_Task, 'duration_secs': 0.01241} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.742035] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.742035] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] dc6523b7-c15c-469f-83a5-5c5dc5f2d230/dc6523b7-c15c-469f-83a5-5c5dc5f2d230.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 970.742257] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-596591b4-3a32-4e98-b587-5c3edb6924b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.763222] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16/8d6c05f2-4bac-41ce-a9f0-4df3d86eba16.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.767312] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-639df7eb-f9ac-46c7-9b82-e44d42367aa2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.783052] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 970.783052] env[61936]: value = "task-1253260" [ 970.783052] env[61936]: _type = "Task" [ 970.783052] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.789068] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 970.789068] env[61936]: value = "task-1253261" [ 970.789068] env[61936]: _type = "Task" [ 970.789068] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.792288] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.801059] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253261, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.841909] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253259, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.948206] env[61936]: DEBUG nova.compute.manager [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-vif-plugged-caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 970.948206] env[61936]: DEBUG oslo_concurrency.lockutils [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.948206] env[61936]: DEBUG oslo_concurrency.lockutils [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.948206] env[61936]: DEBUG oslo_concurrency.lockutils [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.948206] env[61936]: DEBUG nova.compute.manager [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] No waiting events found dispatching network-vif-plugged-caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 970.948206] env[61936]: WARNING nova.compute.manager [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received unexpected event network-vif-plugged-caac4e2a-a9cd-41ca-8d82-9631afe6ed91 for instance with vm_state building and task_state spawning. [ 970.948206] env[61936]: DEBUG nova.compute.manager [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-changed-caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 970.949234] env[61936]: DEBUG nova.compute.manager [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing instance network info cache due to event network-changed-caac4e2a-a9cd-41ca-8d82-9631afe6ed91. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 970.949628] env[61936]: DEBUG oslo_concurrency.lockutils [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.037024] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253256, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.040478] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481fddba-0e2f-4a04-a017-1860a03a09e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.046806] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c4027e-e9d0-4842-87af-29ca4e240535 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.084072] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8529e998-fa8b-4793-b856-9d6560b206b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.096699] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926b72df-9523-4070-a504-2a6345093938 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.120889] env[61936]: DEBUG nova.compute.provider_tree [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.172221] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.172574] env[61936]: DEBUG nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Instance network_info: |[{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 971.173408] env[61936]: DEBUG oslo_concurrency.lockutils [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.173408] env[61936]: DEBUG nova.network.neutron [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing network info cache for port caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.174421] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:57:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'caac4e2a-a9cd-41ca-8d82-9631afe6ed91', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.183919] env[61936]: DEBUG oslo.service.loopingcall [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.185255] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.186189] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45b963db-f44b-45b9-8a08-4c515ae083bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.209612] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.209612] env[61936]: value = "task-1253262" [ 971.209612] env[61936]: _type = "Task" [ 971.209612] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.220592] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253262, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.236919] env[61936]: DEBUG nova.network.neutron [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.238399] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e757ee0-b98e-4bea-8131-560bc6ce9590 tempest-SecurityGroupsTestJSON-1016506914 tempest-SecurityGroupsTestJSON-1016506914-project-member] Lock "f2844173-4114-4158-99ad-ac651efbb607" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.064s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.297250] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253260, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.310623] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253261, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.332935] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253259, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.539420] env[61936]: DEBUG oslo_vmware.api [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253256, 'name': ReconfigVM_Task, 'duration_secs': 1.091773} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.540975] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269943', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'name': 'volume-56048ceb-814a-4561-9098-d35c809208f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377', 'attached_at': '', 'detached_at': '', 'volume_id': '56048ceb-814a-4561-9098-d35c809208f0', 'serial': '56048ceb-814a-4561-9098-d35c809208f0'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 971.590155] env[61936]: DEBUG nova.network.neutron [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updated VIF entry in instance network info cache for port caac4e2a-a9cd-41ca-8d82-9631afe6ed91. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.590488] env[61936]: DEBUG nova.network.neutron [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.623803] env[61936]: DEBUG nova.scheduler.client.report [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 971.720286] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253262, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.740154] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.793975] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709556} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.794261] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] dc6523b7-c15c-469f-83a5-5c5dc5f2d230/dc6523b7-c15c-469f-83a5-5c5dc5f2d230.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 971.794477] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.794726] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0bff5eaf-d43b-4234-aeac-6a04041ac57a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.804951] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253261, 'name': ReconfigVM_Task, 'duration_secs': 0.648996} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.806145] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16/8d6c05f2-4bac-41ce-a9f0-4df3d86eba16.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.806880] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 971.806880] env[61936]: value = "task-1253263" [ 971.806880] env[61936]: _type = "Task" [ 971.806880] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.807078] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be54838b-b5e4-484c-bd62-7436705ca2b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.816816] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253263, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.818066] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 971.818066] env[61936]: value = "task-1253264" [ 971.818066] env[61936]: _type = "Task" [ 971.818066] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.825774] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253264, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.835292] env[61936]: DEBUG oslo_vmware.api [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253259, 'name': PowerOnVM_Task, 'duration_secs': 1.256017} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.835567] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.835757] env[61936]: INFO nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Took 9.91 seconds to spawn the instance on the hypervisor. [ 971.835935] env[61936]: DEBUG nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 971.836747] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9b47aa-a485-4608-98bc-119ee22e0cc8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.093572] env[61936]: DEBUG oslo_concurrency.lockutils [req-89871bfc-a345-4c83-a622-3a5769d6d1f4 req-6fb68674-e309-4e27-b238-364a5b26752a service nova] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.098090] env[61936]: DEBUG nova.objects.instance [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'flavor' on Instance uuid 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.130245] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.438s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.130469] env[61936]: INFO nova.compute.manager [None req-6ec8dff2-e979-4ced-b554-64b7149a63de tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Successfully reverted task state from rebuilding on failure for instance. [ 972.136068] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.542s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.137827] env[61936]: INFO nova.compute.claims [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.192588] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.193523] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e99f0f-8b98-4f24-9993-2891f7604451 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.201187] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.201456] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dfcb89e-4bb4-4b33-a483-024ff104ecf7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.218813] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253262, 'name': CreateVM_Task, 'duration_secs': 0.552059} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.219925] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 972.221251] env[61936]: DEBUG nova.compute.manager [req-ba4f9ce8-fa00-4af9-be24-0aefcd181817 req-602a17a3-e38c-40a3-a2cb-be8002d0a0aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received event network-vif-unplugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 972.221460] env[61936]: DEBUG oslo_concurrency.lockutils [req-ba4f9ce8-fa00-4af9-be24-0aefcd181817 req-602a17a3-e38c-40a3-a2cb-be8002d0a0aa service nova] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.221665] env[61936]: DEBUG oslo_concurrency.lockutils [req-ba4f9ce8-fa00-4af9-be24-0aefcd181817 req-602a17a3-e38c-40a3-a2cb-be8002d0a0aa service nova] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.221837] env[61936]: DEBUG oslo_concurrency.lockutils [req-ba4f9ce8-fa00-4af9-be24-0aefcd181817 req-602a17a3-e38c-40a3-a2cb-be8002d0a0aa service nova] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.222017] env[61936]: DEBUG nova.compute.manager [req-ba4f9ce8-fa00-4af9-be24-0aefcd181817 req-602a17a3-e38c-40a3-a2cb-be8002d0a0aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] No waiting events found dispatching network-vif-unplugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 972.222190] env[61936]: WARNING nova.compute.manager [req-ba4f9ce8-fa00-4af9-be24-0aefcd181817 req-602a17a3-e38c-40a3-a2cb-be8002d0a0aa service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received unexpected event network-vif-unplugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 for instance with vm_state shelved and task_state shelving_offloading. [ 972.222852] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.223016] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.223328] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 972.223948] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7938857d-8633-4f55-a956-1fccf7a0fcb9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.229948] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 972.229948] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b98d37-55df-53da-6eac-50ced46a22ed" [ 972.229948] env[61936]: _type = "Task" [ 972.229948] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.239793] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b98d37-55df-53da-6eac-50ced46a22ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.317711] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253263, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093024} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.321092] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.321092] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc6ecfb-e762-42cb-a66c-24bafa2187d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.327435] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.328746] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.328746] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleting the datastore file [datastore1] 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.329584] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-365d2c28-1176-43ed-86a7-43a6834de872 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.356026] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] dc6523b7-c15c-469f-83a5-5c5dc5f2d230/dc6523b7-c15c-469f-83a5-5c5dc5f2d230.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.356026] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253264, 'name': Rename_Task, 'duration_secs': 0.341972} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.359879] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce4a375e-da59-4dc8-8057-f6d461537534 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.380956] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.386124] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa4a0312-584c-41a1-9e07-f69d19ac65a0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.388329] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 972.388329] env[61936]: value = "task-1253266" [ 972.388329] env[61936]: _type = "Task" [ 972.388329] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.389223] env[61936]: INFO nova.compute.manager [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Took 37.67 seconds to build instance. [ 972.397345] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 972.397345] env[61936]: value = "task-1253268" [ 972.397345] env[61936]: _type = "Task" [ 972.397345] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.397636] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 972.397636] env[61936]: value = "task-1253267" [ 972.397636] env[61936]: _type = "Task" [ 972.397636] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.407731] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.413515] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253268, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.418433] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.744451] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b98d37-55df-53da-6eac-50ced46a22ed, 'name': SearchDatastore_Task, 'duration_secs': 0.009896} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.744451] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.744451] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.744451] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.744451] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.744451] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.744451] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92f6665d-1a56-46aa-8e0e-e05090df4276 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.752560] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.752963] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.753794] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2467cca7-0bda-46eb-9ee6-cc57bda672ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.759139] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 972.759139] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52aeac48-b604-8a89-b76a-e962ce1a6f8f" [ 972.759139] env[61936]: _type = "Task" [ 972.759139] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.766815] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52aeac48-b604-8a89-b76a-e962ce1a6f8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.896025] env[61936]: DEBUG oslo_concurrency.lockutils [None req-32f8fc21-172c-4a50-bd4b-76587eda9f79 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.197s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.908494] env[61936]: DEBUG oslo_vmware.api [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253266, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144119} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.909586] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.909942] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.910259] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.920594] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253267, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.921170] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253268, 'name': PowerOnVM_Task} progress is 76%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.938909] env[61936]: INFO nova.scheduler.client.report [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleted allocations for instance 5526cbd5-b1ad-453b-8401-eee7aa356606 [ 973.110420] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d6dfad9b-2fd2-4aeb-a8cd-f464cd353b34 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.299s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.275275] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52aeac48-b604-8a89-b76a-e962ce1a6f8f, 'name': SearchDatastore_Task, 'duration_secs': 0.010959} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.276823] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57f11ebe-453c-4e98-a8ad-8ee26612f64f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.286176] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 973.286176] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5298f2b9-f418-b8f2-724c-c0c0527d3868" [ 973.286176] env[61936]: _type = "Task" [ 973.286176] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.294634] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5298f2b9-f418-b8f2-724c-c0c0527d3868, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.379511] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8de85d7-2ca0-454b-bc46-50bb857a32d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.388152] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79e81a9-741d-4cdc-8133-334d215f8294 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.429982] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03a8332-059d-4068-9435-180b6237e05d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.438605] env[61936]: DEBUG oslo_vmware.api [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253268, 'name': PowerOnVM_Task, 'duration_secs': 0.949998} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.444882] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.445169] env[61936]: INFO nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Took 9.43 seconds to spawn the instance on the hypervisor. [ 973.445299] env[61936]: DEBUG nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 973.446819] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.446819] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253267, 'name': ReconfigVM_Task, 'duration_secs': 0.846509} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.447578] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5b556e-1e1d-4af1-81da-123044aa1d10 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.450516] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Reconfigured VM instance instance-00000058 to attach disk [datastore2] dc6523b7-c15c-469f-83a5-5c5dc5f2d230/dc6523b7-c15c-469f-83a5-5c5dc5f2d230.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.452099] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0162aa-049a-436a-893f-2cbb2aaf4544 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.456124] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-34e53d9e-30e7-4c94-8333-5ec3ad6792ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.478163] env[61936]: DEBUG nova.compute.provider_tree [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.479555] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 973.479555] env[61936]: value = "task-1253269" [ 973.479555] env[61936]: _type = "Task" [ 973.479555] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.488634] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253269, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.611167] env[61936]: DEBUG nova.compute.manager [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Received event network-changed-746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 973.611167] env[61936]: DEBUG nova.compute.manager [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Refreshing instance network info cache due to event network-changed-746d118b-a889-44a8-a6fe-1f8e1c483c33. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 973.611167] env[61936]: DEBUG oslo_concurrency.lockutils [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] Acquiring lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.611167] env[61936]: DEBUG oslo_concurrency.lockutils [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] Acquired lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.611827] env[61936]: DEBUG nova.network.neutron [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Refreshing network info cache for port 746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.698032] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.698032] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.698032] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.698237] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.698417] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.702528] env[61936]: INFO nova.compute.manager [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Terminating instance [ 973.798122] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5298f2b9-f418-b8f2-724c-c0c0527d3868, 'name': SearchDatastore_Task, 'duration_secs': 0.010222} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.798558] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.798892] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 0137305a-dace-4eda-9d90-7233c001176f/0137305a-dace-4eda-9d90-7233c001176f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.799211] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8b12c39-8516-4b58-802e-a21eeebeae79 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.806092] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 973.806092] env[61936]: value = "task-1253270" [ 973.806092] env[61936]: _type = "Task" [ 973.806092] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.814392] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.986208] env[61936]: DEBUG nova.scheduler.client.report [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 973.996187] env[61936]: INFO nova.compute.manager [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Took 37.53 seconds to build instance. [ 974.005053] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253269, 'name': Rename_Task, 'duration_secs': 0.196957} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.007155] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.007155] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ffd4a3b-e192-438f-92a3-a5bae1c09639 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.013610] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 974.013610] env[61936]: value = "task-1253271" [ 974.013610] env[61936]: _type = "Task" [ 974.013610] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.023273] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253271, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.208054] env[61936]: DEBUG nova.compute.manager [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 974.208395] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.209515] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563a661f-86b3-40dd-b837-8fd7a14d85ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.224284] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.224284] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6aace5c5-b853-416f-9efa-35a800a5cc8a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.228938] env[61936]: DEBUG oslo_vmware.api [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 974.228938] env[61936]: value = "task-1253272" [ 974.228938] env[61936]: _type = "Task" [ 974.228938] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.243963] env[61936]: DEBUG oslo_vmware.api [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.320998] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253270, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469616} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.323112] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 0137305a-dace-4eda-9d90-7233c001176f/0137305a-dace-4eda-9d90-7233c001176f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.323357] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.325080] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-113c33df-d488-4f6b-aa90-ca1b423f286d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.331360] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 974.331360] env[61936]: value = "task-1253273" [ 974.331360] env[61936]: _type = "Task" [ 974.331360] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.345404] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253273, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.351647] env[61936]: DEBUG nova.compute.manager [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received event network-changed-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 974.351845] env[61936]: DEBUG nova.compute.manager [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Refreshing instance network info cache due to event network-changed-1cdd9a15-01e7-40cf-80fc-24bea0b642d3. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 974.352530] env[61936]: DEBUG oslo_concurrency.lockutils [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.352978] env[61936]: DEBUG oslo_concurrency.lockutils [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.352978] env[61936]: DEBUG nova.network.neutron [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Refreshing network info cache for port 1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 974.497412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.361s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.498155] env[61936]: DEBUG nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 974.505027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 15.588s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.505027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0a43a84f-a9c8-4f7c-a50c-b1dbbfeb910e tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.049s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.525122] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253271, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.574018] env[61936]: DEBUG nova.network.neutron [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updated VIF entry in instance network info cache for port 746d118b-a889-44a8-a6fe-1f8e1c483c33. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.574605] env[61936]: DEBUG nova.network.neutron [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating instance_info_cache with network_info: [{"id": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "address": "fa:16:3e:ce:3d:0e", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746d118b-a8", "ovs_interfaceid": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.741803] env[61936]: DEBUG oslo_vmware.api [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253272, 'name': PowerOffVM_Task, 'duration_secs': 0.212746} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.741803] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.741803] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.741803] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7591fa50-53d1-4260-a0d3-2f50561c8602 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.824035] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.824162] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.824767] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleting the datastore file [datastore1] 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.824767] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9b15100-9be8-4d35-87cd-59a10dca895b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.831542] env[61936]: DEBUG oslo_vmware.api [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 974.831542] env[61936]: value = "task-1253275" [ 974.831542] env[61936]: _type = "Task" [ 974.831542] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.843095] env[61936]: DEBUG oslo_vmware.api [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.846420] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253273, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070783} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.846420] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.846651] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93568d6b-2c05-4675-831e-1a47a9ffd9c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.871886] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 0137305a-dace-4eda-9d90-7233c001176f/0137305a-dace-4eda-9d90-7233c001176f.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.872595] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9473afcb-99da-4f9e-9773-2af1a9df4235 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.892973] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 974.892973] env[61936]: value = "task-1253276" [ 974.892973] env[61936]: _type = "Task" [ 974.892973] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.901409] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.964030] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.964280] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.011933] env[61936]: DEBUG nova.compute.utils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.013487] env[61936]: DEBUG nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 975.013659] env[61936]: DEBUG nova.network.neutron [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 975.027159] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253271, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.080773] env[61936]: DEBUG oslo_concurrency.lockutils [req-a10119e3-2198-48f1-957c-e6e270ad744e req-43a06a9f-9a78-4bfc-9730-d9d64538b821 service nova] Releasing lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.090915] env[61936]: DEBUG nova.policy [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b98c1bbe1044663830b18c4ce9e23d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1b7d01ffac74c6fa2306b4839041fdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.249123] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e173b73-c345-470b-a7c6-e64c9a3afd82 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.259851] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7327e691-7960-455e-a6fe-54828978b297 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.267021] env[61936]: DEBUG nova.network.neutron [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updated VIF entry in instance network info cache for port 1cdd9a15-01e7-40cf-80fc-24bea0b642d3. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 975.267021] env[61936]: DEBUG nova.network.neutron [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": null, "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.318134] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4096639-c4e5-4b29-91ad-8b7b3909fb65 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.332712] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded28dee-b45b-42d9-b90a-53fd6b669dd8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.349637] env[61936]: DEBUG nova.compute.provider_tree [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.353896] env[61936]: DEBUG oslo_vmware.api [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253275, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162442} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.354345] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.354533] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.354712] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.354883] env[61936]: INFO nova.compute.manager [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Took 1.15 seconds to destroy the instance on the hypervisor. [ 975.355129] env[61936]: DEBUG oslo.service.loopingcall [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.355324] env[61936]: DEBUG nova.compute.manager [-] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 975.355464] env[61936]: DEBUG nova.network.neutron [-] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 975.403100] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253276, 'name': ReconfigVM_Task, 'duration_secs': 0.314609} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.403392] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 0137305a-dace-4eda-9d90-7233c001176f/0137305a-dace-4eda-9d90-7233c001176f.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.404095] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5171737-3716-4d7d-9d2a-946707e23d95 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.410894] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 975.410894] env[61936]: value = "task-1253277" [ 975.410894] env[61936]: _type = "Task" [ 975.410894] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.421227] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253277, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.466584] env[61936]: DEBUG nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 975.501078] env[61936]: DEBUG nova.network.neutron [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Successfully created port: 18af91fe-34f3-43a1-9f99-73143ee13d3f {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.517622] env[61936]: DEBUG nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 975.529887] env[61936]: DEBUG oslo_vmware.api [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253271, 'name': PowerOnVM_Task, 'duration_secs': 1.132561} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.530166] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.530360] env[61936]: INFO nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Took 9.10 seconds to spawn the instance on the hypervisor. [ 975.530542] env[61936]: DEBUG nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 975.531354] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5461a22-271c-42a4-9d1f-88f3bd938da7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.770188] env[61936]: DEBUG oslo_concurrency.lockutils [req-1cc3d91e-0d54-41a3-9b41-34b8aba36d91 req-71a656db-6c97-4d5a-818d-b9935439b3f2 service nova] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.828969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.858025] env[61936]: DEBUG nova.scheduler.client.report [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 975.931121] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253277, 'name': Rename_Task, 'duration_secs': 0.141256} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.931121] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.931121] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a76c7e5f-4404-414f-8882-3075b7bc07cb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.942762] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 975.942762] env[61936]: value = "task-1253278" [ 975.942762] env[61936]: _type = "Task" [ 975.942762] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.953804] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253278, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.994968] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.058869] env[61936]: INFO nova.compute.manager [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Took 32.06 seconds to build instance. [ 976.460390] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253278, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.534516] env[61936]: DEBUG nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 976.562349] env[61936]: DEBUG oslo_concurrency.lockutils [None req-249dda5b-61e6-45cc-b7bc-c980e9273d86 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.571s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.570400] env[61936]: DEBUG nova.compute.manager [req-61625f0f-8391-4404-b60c-7790ab20ef66 req-66eee40c-5ecc-4dfe-a598-7ef0c76bdccd service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Received event network-vif-deleted-a342f5d6-5a18-4e6f-b046-fd7365684711 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 976.570799] env[61936]: INFO nova.compute.manager [req-61625f0f-8391-4404-b60c-7790ab20ef66 req-66eee40c-5ecc-4dfe-a598-7ef0c76bdccd service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Neutron deleted interface a342f5d6-5a18-4e6f-b046-fd7365684711; detaching it from the instance and deleting it from the info cache [ 976.571101] env[61936]: DEBUG nova.network.neutron [req-61625f0f-8391-4404-b60c-7790ab20ef66 req-66eee40c-5ecc-4dfe-a598-7ef0c76bdccd service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.657590] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e4d9e85ac4f1283175ac03d8acd182af',container_format='bare',created_at=2024-10-10T16:50:02Z,direct_url=,disk_format='vmdk',id=af56f59f-c70a-4b04-91e5-d2e24b68120b,min_disk=1,min_ram=0,name='tempest-test-snap-1300946831',owner='e1b7d01ffac74c6fa2306b4839041fdd',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T16:50:17Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 976.657590] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.657590] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 976.657590] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.657851] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 976.657894] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 976.658415] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 976.658681] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 976.658897] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 976.659092] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 976.659291] env[61936]: DEBUG nova.virt.hardware [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 976.660633] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5162caf4-db8f-4bb1-9535-535f22fb740a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.669860] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c531525f-b40a-439b-b565-52870a796e52 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.872248] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.370s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.879348] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.014s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.879348] env[61936]: DEBUG nova.objects.instance [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lazy-loading 'resources' on Instance uuid 1a12fd59-bacd-42b3-99c1-8910f349e2cb {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.956378] env[61936]: DEBUG oslo_vmware.api [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253278, 'name': PowerOnVM_Task, 'duration_secs': 0.549221} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.956486] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.958913] env[61936]: INFO nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Took 8.13 seconds to spawn the instance on the hypervisor. [ 976.958913] env[61936]: DEBUG nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 976.958913] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d72e390-f11b-4ba4-9611-d2ece58e2ec9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.974788] env[61936]: DEBUG nova.network.neutron [-] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.077769] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aecb2127-bce2-4e7b-b3a1-a91c2f33943c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.088125] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6114d3cf-883e-4bcb-a58e-cfb925a11464 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.123093] env[61936]: DEBUG nova.compute.manager [req-61625f0f-8391-4404-b60c-7790ab20ef66 req-66eee40c-5ecc-4dfe-a598-7ef0c76bdccd service nova] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Detach interface failed, port_id=a342f5d6-5a18-4e6f-b046-fd7365684711, reason: Instance 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 977.452857] env[61936]: INFO nova.scheduler.client.report [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted allocation for migration 43339264-3f33-4305-8859-baeb09cac628 [ 977.481484] env[61936]: INFO nova.compute.manager [-] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Took 2.12 seconds to deallocate network for instance. [ 977.481484] env[61936]: INFO nova.compute.manager [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Took 30.49 seconds to build instance. [ 977.487543] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.487992] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.488076] env[61936]: INFO nova.compute.manager [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Shelving [ 977.695155] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc65313d-ea65-42d1-adfb-446bb543c256 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.703201] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4ec9ca-c49f-48b3-9699-2590458ad065 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.749680] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c77560-acc4-4c3b-ac14-e3a683eaa2e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.758453] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73693a6e-abca-45c4-ad0a-cd16e8a5dd0b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.778977] env[61936]: DEBUG nova.compute.provider_tree [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.900616] env[61936]: DEBUG nova.network.neutron [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Successfully updated port: 18af91fe-34f3-43a1-9f99-73143ee13d3f {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.957985] env[61936]: DEBUG oslo_concurrency.lockutils [None req-30d66d24-ff79-4d4e-b8c0-cca8a96d7282 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 22.618s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.985607] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14f62cf9-1205-4f5a-b3b6-1f9319932954 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "0137305a-dace-4eda-9d90-7233c001176f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.015s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.991130] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.283032] env[61936]: DEBUG nova.scheduler.client.report [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 978.403508] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "refresh_cache-62b33dd9-e523-4309-94e6-59671fcfbaea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.403794] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "refresh_cache-62b33dd9-e523-4309-94e6-59671fcfbaea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.404044] env[61936]: DEBUG nova.network.neutron [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.469329] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.469613] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.505182] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.505538] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d633f5ce-f1cf-4048-a8c5-e08df8e17f17 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.517031] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 978.517031] env[61936]: value = "task-1253279" [ 978.517031] env[61936]: _type = "Task" [ 978.517031] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.522849] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.791026] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.798636] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.333s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.800658] env[61936]: INFO nova.compute.claims [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.809707] env[61936]: DEBUG nova.compute.manager [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Received event network-vif-plugged-18af91fe-34f3-43a1-9f99-73143ee13d3f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 978.809940] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] Acquiring lock "62b33dd9-e523-4309-94e6-59671fcfbaea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.810898] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.810898] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.811029] env[61936]: DEBUG nova.compute.manager [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] No waiting events found dispatching network-vif-plugged-18af91fe-34f3-43a1-9f99-73143ee13d3f {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 978.811376] env[61936]: WARNING nova.compute.manager [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Received unexpected event network-vif-plugged-18af91fe-34f3-43a1-9f99-73143ee13d3f for instance with vm_state building and task_state spawning. [ 978.811627] env[61936]: DEBUG nova.compute.manager [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Received event network-changed-18af91fe-34f3-43a1-9f99-73143ee13d3f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 978.811801] env[61936]: DEBUG nova.compute.manager [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Refreshing instance network info cache due to event network-changed-18af91fe-34f3-43a1-9f99-73143ee13d3f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 978.811983] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] Acquiring lock "refresh_cache-62b33dd9-e523-4309-94e6-59671fcfbaea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.964110] env[61936]: DEBUG nova.network.neutron [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 978.972418] env[61936]: DEBUG nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 979.036174] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253279, 'name': PowerOffVM_Task, 'duration_secs': 0.353378} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.036580] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.037477] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e847634-a258-4815-8cf2-c3129c3d1f61 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.060693] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5685f3-6658-4f90-b22c-84fdb856e529 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.220890] env[61936]: DEBUG nova.network.neutron [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Updating instance_info_cache with network_info: [{"id": "18af91fe-34f3-43a1-9f99-73143ee13d3f", "address": "fa:16:3e:24:3d:96", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18af91fe-34", "ovs_interfaceid": "18af91fe-34f3-43a1-9f99-73143ee13d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.321830] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f232d32f-c237-423e-9754-a76c86093570 tempest-ServerActionsV293TestJSON-984373365 tempest-ServerActionsV293TestJSON-984373365-project-member] Lock "1a12fd59-bacd-42b3-99c1-8910f349e2cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.046s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.500434] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.573222] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Creating Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 979.573649] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c6fda990-bd37-4e12-a80e-408dc8003615 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.583354] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 979.583354] env[61936]: value = "task-1253280" [ 979.583354] env[61936]: _type = "Task" [ 979.583354] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.592599] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253280, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.722969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "refresh_cache-62b33dd9-e523-4309-94e6-59671fcfbaea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.723434] env[61936]: DEBUG nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Instance network_info: |[{"id": "18af91fe-34f3-43a1-9f99-73143ee13d3f", "address": "fa:16:3e:24:3d:96", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18af91fe-34", "ovs_interfaceid": "18af91fe-34f3-43a1-9f99-73143ee13d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 979.723713] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] Acquired lock "refresh_cache-62b33dd9-e523-4309-94e6-59671fcfbaea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.723974] env[61936]: DEBUG nova.network.neutron [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Refreshing network info cache for port 18af91fe-34f3-43a1-9f99-73143ee13d3f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.725398] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:3d:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18af91fe-34f3-43a1-9f99-73143ee13d3f', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.735084] env[61936]: DEBUG oslo.service.loopingcall [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.741579] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 979.742216] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc752bcd-258f-4f22-82ca-109d9f73cf88 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.777662] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.777662] env[61936]: value = "task-1253281" [ 979.777662] env[61936]: _type = "Task" [ 979.777662] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.786273] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253281, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.100618] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253280, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.118649] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b91012e-4151-467b-aed1-298204044f1a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.125998] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534c2f15-98ea-4723-94ba-788ed14b2692 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.159235] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fac08de-5ce8-4be6-a476-a0d54af3a7a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.166358] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16b341f-6bc0-4f87-acc9-6d0ff4d4e9c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.171011] env[61936]: DEBUG nova.network.neutron [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Updated VIF entry in instance network info cache for port 18af91fe-34f3-43a1-9f99-73143ee13d3f. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.171361] env[61936]: DEBUG nova.network.neutron [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Updating instance_info_cache with network_info: [{"id": "18af91fe-34f3-43a1-9f99-73143ee13d3f", "address": "fa:16:3e:24:3d:96", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18af91fe-34", "ovs_interfaceid": "18af91fe-34f3-43a1-9f99-73143ee13d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.187178] env[61936]: DEBUG nova.compute.provider_tree [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.288704] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253281, 'name': CreateVM_Task, 'duration_secs': 0.339047} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.288879] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 980.290283] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.290492] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.290881] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 980.291168] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b1ee5e0-4e6f-4697-843b-ee59015807b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.295795] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 980.295795] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523be66b-ef9e-8460-56e4-48c51c0df9b6" [ 980.295795] env[61936]: _type = "Task" [ 980.295795] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.303966] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523be66b-ef9e-8460-56e4-48c51c0df9b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.339856] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.340174] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.340390] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.340575] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.340803] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.343301] env[61936]: INFO nova.compute.manager [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Terminating instance [ 980.594485] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253280, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.676975] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a4d4900-58a7-42b5-80a4-5bd888e15c60 req-f5d22fad-519e-486f-a8ef-d1622370e98f service nova] Releasing lock "refresh_cache-62b33dd9-e523-4309-94e6-59671fcfbaea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.690064] env[61936]: DEBUG nova.scheduler.client.report [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 980.806522] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.806815] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Processing image af56f59f-c70a-4b04-91e5-d2e24b68120b {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.807126] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b/af56f59f-c70a-4b04-91e5-d2e24b68120b.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.807230] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b/af56f59f-c70a-4b04-91e5-d2e24b68120b.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.807394] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.807699] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d11bc54-3505-426d-9cba-1a12c581014d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.818074] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.818283] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 980.819152] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2caca48-12d6-4dd2-b6d4-2a3bda796be4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.825271] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 980.825271] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f1ff93-b697-e81c-cc09-7c0bc67bff8f" [ 980.825271] env[61936]: _type = "Task" [ 980.825271] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.833031] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f1ff93-b697-e81c-cc09-7c0bc67bff8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.842672] env[61936]: DEBUG nova.compute.manager [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-changed-caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 980.842863] env[61936]: DEBUG nova.compute.manager [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing instance network info cache due to event network-changed-caac4e2a-a9cd-41ca-8d82-9631afe6ed91. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 980.843079] env[61936]: DEBUG oslo_concurrency.lockutils [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.843225] env[61936]: DEBUG oslo_concurrency.lockutils [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.843382] env[61936]: DEBUG nova.network.neutron [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing network info cache for port caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 980.847252] env[61936]: DEBUG nova.compute.manager [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 980.847455] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.848332] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c032a6-ca3a-40d6-b1aa-d9cfebd84b8f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.856173] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.856898] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee4c0437-c0e8-4565-aad4-f15453319c49 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.865703] env[61936]: DEBUG oslo_vmware.api [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 980.865703] env[61936]: value = "task-1253282" [ 980.865703] env[61936]: _type = "Task" [ 980.865703] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.874742] env[61936]: DEBUG oslo_vmware.api [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253282, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.981134] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.981748] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.096040] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253280, 'name': CreateSnapshot_Task, 'duration_secs': 1.061394} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.096321] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Created Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 981.097252] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3eb89d-88dc-4f08-87ff-655cda945480 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.194615] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.195259] env[61936]: DEBUG nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 981.197956] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.497s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.198234] env[61936]: DEBUG nova.objects.instance [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lazy-loading 'resources' on Instance uuid 45fd1d4d-7125-43a7-afe7-53f737fb9f8e {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.335029] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Preparing fetch location {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 981.335872] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Fetch image to [datastore1] OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f/OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f.vmdk {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 981.335872] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Downloading stream optimized image af56f59f-c70a-4b04-91e5-d2e24b68120b to [datastore1] OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f/OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f.vmdk on the data store datastore1 as vApp {{(pid=61936) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 981.336079] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Downloading image file data af56f59f-c70a-4b04-91e5-d2e24b68120b to the ESX as VM named 'OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f' {{(pid=61936) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 981.394656] env[61936]: DEBUG oslo_vmware.api [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253282, 'name': PowerOffVM_Task, 'duration_secs': 0.435042} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.395057] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.395248] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.395528] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66e57150-a0ae-48c2-8ee7-02f80af96009 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.425041] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 981.425041] env[61936]: value = "resgroup-9" [ 981.425041] env[61936]: _type = "ResourcePool" [ 981.425041] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 981.425400] env[61936]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d9ecc9ad-60f9-4a50-b3d8-e23644d2d2f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.446947] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lease: (returnval){ [ 981.446947] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52696142-d6b2-00d8-4554-1a15e73252cc" [ 981.446947] env[61936]: _type = "HttpNfcLease" [ 981.446947] env[61936]: } obtained for vApp import into resource pool (val){ [ 981.446947] env[61936]: value = "resgroup-9" [ 981.446947] env[61936]: _type = "ResourcePool" [ 981.446947] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 981.447428] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the lease: (returnval){ [ 981.447428] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52696142-d6b2-00d8-4554-1a15e73252cc" [ 981.447428] env[61936]: _type = "HttpNfcLease" [ 981.447428] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 981.456090] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 981.456090] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52696142-d6b2-00d8-4554-1a15e73252cc" [ 981.456090] env[61936]: _type = "HttpNfcLease" [ 981.456090] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 981.476326] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.476645] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.476844] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleting the datastore file [datastore2] 9bd97f84-7c21-44ad-9ed0-d7e4097e264e {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.477148] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8ecda0a-f31b-401a-be26-5f239ffe16fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.483915] env[61936]: DEBUG oslo_vmware.api [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 981.483915] env[61936]: value = "task-1253285" [ 981.483915] env[61936]: _type = "Task" [ 981.483915] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.484377] env[61936]: DEBUG nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 981.501807] env[61936]: DEBUG oslo_vmware.api [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253285, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.614475] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Creating linked-clone VM from snapshot {{(pid=61936) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 981.614854] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a788d1b5-9410-4b6a-80c0-0976f40b3985 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.618803] env[61936]: DEBUG nova.network.neutron [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updated VIF entry in instance network info cache for port caac4e2a-a9cd-41ca-8d82-9631afe6ed91. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 981.619139] env[61936]: DEBUG nova.network.neutron [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.624990] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 981.624990] env[61936]: value = "task-1253286" [ 981.624990] env[61936]: _type = "Task" [ 981.624990] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.632709] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253286, 'name': CloneVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.701805] env[61936]: DEBUG nova.compute.utils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.705952] env[61936]: DEBUG nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 981.706450] env[61936]: DEBUG nova.network.neutron [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.751478] env[61936]: DEBUG nova.policy [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9bb41ed24f244345b83f2a16f4f10804', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f1b73bc1dea4df0a995655f3bd50d14', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.955158] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 981.955158] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52696142-d6b2-00d8-4554-1a15e73252cc" [ 981.955158] env[61936]: _type = "HttpNfcLease" [ 981.955158] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 981.957771] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c62e049-8887-439a-883e-0895c191b3c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.965252] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7abe3e-9ba0-40b8-a7bb-127ee481cbb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.001874] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89979a3a-257d-488e-8e4f-9fc51330e347 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.011788] env[61936]: DEBUG oslo_vmware.api [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253285, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.233455} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.014367] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 982.014594] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 982.014818] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 982.014902] env[61936]: INFO nova.compute.manager [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 982.015166] env[61936]: DEBUG oslo.service.loopingcall [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.016124] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.016469] env[61936]: DEBUG nova.compute.manager [-] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 982.016469] env[61936]: DEBUG nova.network.neutron [-] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 982.019202] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df96802-cb71-420d-acde-23710107bfba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.034830] env[61936]: DEBUG nova.compute.provider_tree [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.122207] env[61936]: DEBUG oslo_concurrency.lockutils [req-6b68120b-ce6c-4454-bd00-e9766157a340 req-df9d3187-1c9c-4aea-a986-b48e9755372c service nova] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.134999] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253286, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.207411] env[61936]: DEBUG nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 982.239740] env[61936]: DEBUG nova.network.neutron [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Successfully created port: 7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 982.456210] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 982.456210] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52696142-d6b2-00d8-4554-1a15e73252cc" [ 982.456210] env[61936]: _type = "HttpNfcLease" [ 982.456210] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 982.456542] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 982.456542] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52696142-d6b2-00d8-4554-1a15e73252cc" [ 982.456542] env[61936]: _type = "HttpNfcLease" [ 982.456542] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 982.457272] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f410950c-dc29-451a-95cb-f0b02114f084 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.467271] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa1978-d0e4-1c4f-2b29-ac2d5ecb86ea/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 982.467457] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa1978-d0e4-1c4f-2b29-ac2d5ecb86ea/disk-0.vmdk. {{(pid=61936) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 982.537340] env[61936]: DEBUG nova.scheduler.client.report [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 982.544137] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0bd85957-1cd3-48fa-9e2c-1b0883dbdb7b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.635307] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253286, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.875082] env[61936]: DEBUG nova.compute.manager [req-afaac962-57bc-4dd4-a9ec-424bec3fc021 req-a2e8cf8e-9fa4-4961-9772-cd4483eeb5b6 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Received event network-vif-deleted-432e9691-4e6f-4550-a7f4-7396d05e96d7 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 982.875373] env[61936]: INFO nova.compute.manager [req-afaac962-57bc-4dd4-a9ec-424bec3fc021 req-a2e8cf8e-9fa4-4961-9772-cd4483eeb5b6 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Neutron deleted interface 432e9691-4e6f-4550-a7f4-7396d05e96d7; detaching it from the instance and deleting it from the info cache [ 982.875565] env[61936]: DEBUG nova.network.neutron [req-afaac962-57bc-4dd4-a9ec-424bec3fc021 req-a2e8cf8e-9fa4-4961-9772-cd4483eeb5b6 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.046131] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.051174] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.029s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.051430] env[61936]: DEBUG nova.objects.instance [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lazy-loading 'resources' on Instance uuid a99a6fef-3cef-409a-b001-4aca97f852c0 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.073859] env[61936]: INFO nova.scheduler.client.report [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted allocations for instance 45fd1d4d-7125-43a7-afe7-53f737fb9f8e [ 983.092541] env[61936]: DEBUG nova.network.neutron [-] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.136113] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253286, 'name': CloneVM_Task} progress is 95%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.217056] env[61936]: DEBUG nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 983.242338] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Completed reading data from the image iterator. {{(pid=61936) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 983.242569] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa1978-d0e4-1c4f-2b29-ac2d5ecb86ea/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 983.243642] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebc6a84-9325-4a2f-ba9a-501c9742ba67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.248281] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 983.248556] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.248757] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 983.248950] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.249113] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 983.249264] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 983.249520] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 983.249707] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 983.249878] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 983.250051] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 983.250228] env[61936]: DEBUG nova.virt.hardware [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 983.251008] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56cec821-e84a-4e99-95a2-f1820c28fd85 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.257014] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa1978-d0e4-1c4f-2b29-ac2d5ecb86ea/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 983.257190] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa1978-d0e4-1c4f-2b29-ac2d5ecb86ea/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 983.259301] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-6df21bb0-0975-4812-8cfd-07ed60db35f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.261955] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bb9059-18e9-4fe1-8d7f-a4b5762e609f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.378925] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d74e305e-5194-404a-9d21-8906a40524a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.388955] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213b5c06-aa05-42f8-9e4a-556b5226577b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.424685] env[61936]: DEBUG nova.compute.manager [req-afaac962-57bc-4dd4-a9ec-424bec3fc021 req-a2e8cf8e-9fa4-4961-9772-cd4483eeb5b6 service nova] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Detach interface failed, port_id=432e9691-4e6f-4550-a7f4-7396d05e96d7, reason: Instance 9bd97f84-7c21-44ad-9ed0-d7e4097e264e could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 983.454251] env[61936]: DEBUG oslo_vmware.rw_handles [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fa1978-d0e4-1c4f-2b29-ac2d5ecb86ea/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 983.454526] env[61936]: INFO nova.virt.vmwareapi.images [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Downloaded image file data af56f59f-c70a-4b04-91e5-d2e24b68120b [ 983.455480] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b6158c-2afe-4997-b685-1c9bc706073b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.472739] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0834d86f-613b-4a08-98cc-0cc4354f4413 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.499519] env[61936]: INFO nova.virt.vmwareapi.images [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] The imported VM was unregistered [ 983.501919] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Caching image {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 983.502172] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating directory with path [datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.502456] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fccdac95-8141-4f6a-83b4-0d4cfa89e23f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.536495] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Created directory with path [datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.536719] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f/OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f.vmdk to [datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b/af56f59f-c70a-4b04-91e5-d2e24b68120b.vmdk. {{(pid=61936) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 983.536996] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-cc82407a-3576-48b2-a70e-456a5d972bfb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.544774] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 983.544774] env[61936]: value = "task-1253288" [ 983.544774] env[61936]: _type = "Task" [ 983.544774] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.553341] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253288, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.582942] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b9f98c19-4af2-40d7-adb3-cd6c22a27acb tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "45fd1d4d-7125-43a7-afe7-53f737fb9f8e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.480s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.594681] env[61936]: INFO nova.compute.manager [-] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Took 1.58 seconds to deallocate network for instance. [ 983.640208] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253286, 'name': CloneVM_Task, 'duration_secs': 1.846454} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.640208] env[61936]: INFO nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Created linked-clone VM from snapshot [ 983.640857] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a9bfd2-96e2-45ac-929d-60da08dd83a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.649695] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Uploading image e8676683-128d-4455-aff8-ca32f97ce13e {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 983.668953] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 983.668953] env[61936]: value = "vm-270005" [ 983.668953] env[61936]: _type = "VirtualMachine" [ 983.668953] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 983.669512] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-611a35d0-b36b-4437-8968-fad92df2257f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.677724] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lease: (returnval){ [ 983.677724] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52526d68-1860-2220-c6ef-2448cc8f18d2" [ 983.677724] env[61936]: _type = "HttpNfcLease" [ 983.677724] env[61936]: } obtained for exporting VM: (result){ [ 983.677724] env[61936]: value = "vm-270005" [ 983.677724] env[61936]: _type = "VirtualMachine" [ 983.677724] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 983.678106] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the lease: (returnval){ [ 983.678106] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52526d68-1860-2220-c6ef-2448cc8f18d2" [ 983.678106] env[61936]: _type = "HttpNfcLease" [ 983.678106] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 983.685009] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 983.685009] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52526d68-1860-2220-c6ef-2448cc8f18d2" [ 983.685009] env[61936]: _type = "HttpNfcLease" [ 983.685009] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 983.777513] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8524006a-0cc7-471d-a8fa-4cdec8e330e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.784760] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac649e57-d2e8-489b-9ec5-200aaaf0469e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.817331] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca112fd-fe1b-4bbe-b855-30ca480f6ce1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.824479] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d251b4e-a117-4ea1-8458-c5767413b95f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.837553] env[61936]: DEBUG nova.compute.provider_tree [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.056225] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253288, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.057069] env[61936]: DEBUG nova.network.neutron [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Successfully updated port: 7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.100871] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.187676] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 984.187676] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52526d68-1860-2220-c6ef-2448cc8f18d2" [ 984.187676] env[61936]: _type = "HttpNfcLease" [ 984.187676] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 984.187961] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 984.187961] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52526d68-1860-2220-c6ef-2448cc8f18d2" [ 984.187961] env[61936]: _type = "HttpNfcLease" [ 984.187961] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 984.188825] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742e870b-ba9f-4664-abe6-efbdfd85a496 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.197498] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52de195b-1584-dbcb-9d9c-f9975ea22015/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 984.197776] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52de195b-1584-dbcb-9d9c-f9975ea22015/disk-0.vmdk for reading. {{(pid=61936) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 984.301244] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-69d36c91-cd6e-4861-9031-4323305538f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.340904] env[61936]: DEBUG nova.scheduler.client.report [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 984.555224] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253288, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.560225] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.560362] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.560495] env[61936]: DEBUG nova.network.neutron [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.846296] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.795s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.849237] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.403s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.849510] env[61936]: DEBUG nova.objects.instance [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lazy-loading 'resources' on Instance uuid 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.873100] env[61936]: INFO nova.scheduler.client.report [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Deleted allocations for instance a99a6fef-3cef-409a-b001-4aca97f852c0 [ 984.908077] env[61936]: DEBUG nova.compute.manager [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Received event network-vif-plugged-7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 984.909191] env[61936]: DEBUG oslo_concurrency.lockutils [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] Acquiring lock "484e5d41-29f4-4845-9633-157c03766978-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.909191] env[61936]: DEBUG oslo_concurrency.lockutils [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] Lock "484e5d41-29f4-4845-9633-157c03766978-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.909293] env[61936]: DEBUG oslo_concurrency.lockutils [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] Lock "484e5d41-29f4-4845-9633-157c03766978-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.909399] env[61936]: DEBUG nova.compute.manager [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] No waiting events found dispatching network-vif-plugged-7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 984.909659] env[61936]: WARNING nova.compute.manager [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Received unexpected event network-vif-plugged-7450de29-76d5-40b0-ae76-a79b3455a9bc for instance with vm_state building and task_state spawning. [ 984.909872] env[61936]: DEBUG nova.compute.manager [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Received event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 984.910052] env[61936]: DEBUG nova.compute.manager [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing instance network info cache due to event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 984.910228] env[61936]: DEBUG oslo_concurrency.lockutils [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] Acquiring lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.056883] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253288, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.095499] env[61936]: DEBUG nova.network.neutron [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 985.257928] env[61936]: DEBUG nova.network.neutron [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.354506] env[61936]: DEBUG nova.objects.instance [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lazy-loading 'numa_topology' on Instance uuid 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.381714] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c92a754f-c207-4a91-a80d-0e8896d770fa tempest-MultipleCreateTestJSON-834629491 tempest-MultipleCreateTestJSON-834629491-project-member] Lock "a99a6fef-3cef-409a-b001-4aca97f852c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.967s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.558264] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253288, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.760788] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.761148] env[61936]: DEBUG nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Instance network_info: |[{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 985.761490] env[61936]: DEBUG oslo_concurrency.lockutils [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] Acquired lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.761668] env[61936]: DEBUG nova.network.neutron [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.762943] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:ea:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7450de29-76d5-40b0-ae76-a79b3455a9bc', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.770782] env[61936]: DEBUG oslo.service.loopingcall [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.773981] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 484e5d41-29f4-4845-9633-157c03766978] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.774562] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da91a086-6da4-4907-a88a-24c453bc3eae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.797640] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.797640] env[61936]: value = "task-1253290" [ 985.797640] env[61936]: _type = "Task" [ 985.797640] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.808270] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253290, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.857932] env[61936]: DEBUG nova.objects.base [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Object Instance<5526cbd5-b1ad-453b-8401-eee7aa356606> lazy-loaded attributes: resources,numa_topology {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 986.061688] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253288, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.125424] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f685dde5-324f-4238-8ea5-3c60c1dac3d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.133347] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9085340-f665-4148-8dc0-66e58e259cb8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.174584] env[61936]: DEBUG nova.network.neutron [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updated VIF entry in instance network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.174984] env[61936]: DEBUG nova.network.neutron [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.176831] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fdd962-cf12-4bb2-a0f6-ccd029123191 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.186026] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717294cb-0b69-4187-bee2-6e9ed14b580d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.200734] env[61936]: DEBUG nova.compute.provider_tree [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.309466] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253290, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.560041] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253288, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.668663} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.560210] env[61936]: INFO nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f/OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f.vmdk to [datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b/af56f59f-c70a-4b04-91e5-d2e24b68120b.vmdk. [ 986.560636] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Cleaning up location [datastore1] OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 986.560636] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_4923ffba-6582-49f9-81c9-34b1b7825c6f {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.561039] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a5eab36-da87-476f-99e2-d3fc44dc5940 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.569095] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 986.569095] env[61936]: value = "task-1253291" [ 986.569095] env[61936]: _type = "Task" [ 986.569095] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.578393] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.681338] env[61936]: DEBUG oslo_concurrency.lockutils [req-e7f7210a-b55d-480b-b3ba-31556993fe57 req-8eea48d6-4be4-433b-9140-c1f619a42e04 service nova] Releasing lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.704874] env[61936]: DEBUG nova.scheduler.client.report [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 986.812000] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253290, 'name': CreateVM_Task, 'duration_secs': 0.664825} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.812246] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 484e5d41-29f4-4845-9633-157c03766978] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.813117] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.813343] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.813728] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.814055] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f0caf75-de88-4adc-be99-a31238eebe25 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.820042] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 986.820042] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52571141-e933-fdd8-fda7-8ea4f8544bec" [ 986.820042] env[61936]: _type = "Task" [ 986.820042] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.830026] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52571141-e933-fdd8-fda7-8ea4f8544bec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.079593] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041948} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.079891] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.080050] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b/af56f59f-c70a-4b04-91e5-d2e24b68120b.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.080307] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b/af56f59f-c70a-4b04-91e5-d2e24b68120b.vmdk to [datastore1] 62b33dd9-e523-4309-94e6-59671fcfbaea/62b33dd9-e523-4309-94e6-59671fcfbaea.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.080604] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39c8ad9c-25af-4dfb-9352-c9d8d1ec2ca3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.087162] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 987.087162] env[61936]: value = "task-1253292" [ 987.087162] env[61936]: _type = "Task" [ 987.087162] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.095163] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253292, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.210023] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.360s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.212078] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.217s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.213630] env[61936]: INFO nova.compute.claims [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.338240] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52571141-e933-fdd8-fda7-8ea4f8544bec, 'name': SearchDatastore_Task, 'duration_secs': 0.009798} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.338694] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.338865] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.339135] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.339288] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.339477] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.339822] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b710cf18-be77-4de3-bb9c-7605ace3ef4f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.358395] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.358777] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.359929] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-790f9e14-1276-444f-8222-5ff9ec9016fe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.367205] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 987.367205] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c98204-0e39-37cf-6c4a-2b830da22211" [ 987.367205] env[61936]: _type = "Task" [ 987.367205] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.379527] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c98204-0e39-37cf-6c4a-2b830da22211, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.598154] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253292, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.723478] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6f700da6-cabc-493a-ab60-cb3434650d33 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 33.908s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.724490] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 11.896s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.724681] env[61936]: INFO nova.compute.manager [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Unshelving [ 987.879188] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c98204-0e39-37cf-6c4a-2b830da22211, 'name': SearchDatastore_Task, 'duration_secs': 0.077816} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.880192] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db5f0d48-53cb-4f1a-a069-95494c3f0e57 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.886571] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 987.886571] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527c0a88-da76-46b5-ee0e-252a4e531d27" [ 987.886571] env[61936]: _type = "Task" [ 987.886571] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.894930] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527c0a88-da76-46b5-ee0e-252a4e531d27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.097938] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253292, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.401859] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527c0a88-da76-46b5-ee0e-252a4e531d27, 'name': SearchDatastore_Task, 'duration_secs': 0.085235} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.402284] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.402608] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 484e5d41-29f4-4845-9633-157c03766978/484e5d41-29f4-4845-9633-157c03766978.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 988.402972] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ba389ba-07da-4268-87a6-f943f906eae7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.413441] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 988.413441] env[61936]: value = "task-1253293" [ 988.413441] env[61936]: _type = "Task" [ 988.413441] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.424234] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.460450] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e409abcb-7005-4ddb-8b88-558d59bde1d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.467665] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb1b3b1-720b-4faa-983d-422110d570c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.500143] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426c4af2-2714-475b-a0ab-415b38a7dd1e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.508032] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb931a57-f8e0-4e73-8975-a2214e52b28b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.521799] env[61936]: DEBUG nova.compute.provider_tree [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.599859] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253292, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.755392] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.924810] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.027044] env[61936]: DEBUG nova.scheduler.client.report [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 989.099752] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253292, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.425943] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253293, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.532051] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.320s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.532640] env[61936]: DEBUG nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 989.536522] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.545s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.537670] env[61936]: DEBUG nova.objects.instance [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'resources' on Instance uuid 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.601226] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253292, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.28925} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.601734] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/af56f59f-c70a-4b04-91e5-d2e24b68120b/af56f59f-c70a-4b04-91e5-d2e24b68120b.vmdk to [datastore1] 62b33dd9-e523-4309-94e6-59671fcfbaea/62b33dd9-e523-4309-94e6-59671fcfbaea.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.603040] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd884787-e983-49fc-a86c-75c648627aa6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.628485] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 62b33dd9-e523-4309-94e6-59671fcfbaea/62b33dd9-e523-4309-94e6-59671fcfbaea.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.629206] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f18fc956-c8a8-4f50-aca9-40810ce8be92 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.649258] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 989.649258] env[61936]: value = "task-1253294" [ 989.649258] env[61936]: _type = "Task" [ 989.649258] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.659637] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253294, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.927408] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253293, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.439321} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.927408] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 484e5d41-29f4-4845-9633-157c03766978/484e5d41-29f4-4845-9633-157c03766978.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.927408] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.927408] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf0f7f90-5653-4821-a993-7424617ffc15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.935525] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 989.935525] env[61936]: value = "task-1253295" [ 989.935525] env[61936]: _type = "Task" [ 989.935525] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.946011] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253295, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.037346] env[61936]: DEBUG nova.compute.utils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 990.038884] env[61936]: DEBUG nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 990.039271] env[61936]: DEBUG nova.network.neutron [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 990.112526] env[61936]: DEBUG nova.policy [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b05bdcace008487dba580405eb672ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90c29449db6c4beaa7f190d9225cb08a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 990.161784] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253294, 'name': ReconfigVM_Task, 'duration_secs': 0.32907} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.161784] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 62b33dd9-e523-4309-94e6-59671fcfbaea/62b33dd9-e523-4309-94e6-59671fcfbaea.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 990.162356] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f762776-ad5d-4366-9844-2297acbba1fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.169368] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 990.169368] env[61936]: value = "task-1253296" [ 990.169368] env[61936]: _type = "Task" [ 990.169368] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.178683] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253296, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.279928] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7766875d-ef1d-49a1-b836-c0f597a15c83 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.288487] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49f6fdc-2b1c-46c0-b85e-92b15ff9b19f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.320604] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c213e8-d53e-4c5d-9d4a-995977ae7ea8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.328825] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bd7fd7-878e-496f-9850-413f57ba47fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.344317] env[61936]: DEBUG nova.compute.provider_tree [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.453284] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253295, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066163} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.454033] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.456097] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd327522-ee38-4967-b60d-dcc3054ab867 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.482453] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 484e5d41-29f4-4845-9633-157c03766978/484e5d41-29f4-4845-9633-157c03766978.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.482918] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4a58615-71f8-493b-8a14-4cd5730ce5c8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.506825] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 990.506825] env[61936]: value = "task-1253297" [ 990.506825] env[61936]: _type = "Task" [ 990.506825] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.518824] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253297, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.545120] env[61936]: DEBUG nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 990.680819] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253296, 'name': Rename_Task, 'duration_secs': 0.140551} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.681158] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.681414] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-145f76a7-eb7b-479c-8afc-e56b51237294 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.690405] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 990.690405] env[61936]: value = "task-1253298" [ 990.690405] env[61936]: _type = "Task" [ 990.690405] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.700283] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253298, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.710386] env[61936]: DEBUG nova.network.neutron [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Successfully created port: e21d453a-d68d-4497-9042-ebc1a0dbec83 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.847709] env[61936]: DEBUG nova.scheduler.client.report [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 991.020236] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253297, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.201940] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253298, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.354232] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.818s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.357184] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.857s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.359369] env[61936]: INFO nova.compute.claims [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.385893] env[61936]: INFO nova.scheduler.client.report [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted allocations for instance 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377 [ 991.518308] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253297, 'name': ReconfigVM_Task, 'duration_secs': 0.522295} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.518966] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 484e5d41-29f4-4845-9633-157c03766978/484e5d41-29f4-4845-9633-157c03766978.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.519585] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c544f38f-d461-4532-9164-31cb055b625d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.526301] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 991.526301] env[61936]: value = "task-1253299" [ 991.526301] env[61936]: _type = "Task" [ 991.526301] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.536850] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253299, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.555707] env[61936]: DEBUG nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 991.590051] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 991.590332] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.590649] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 991.591022] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.591326] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 991.591668] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 991.592042] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 991.592361] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 991.592674] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 991.593427] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 991.593427] env[61936]: DEBUG nova.virt.hardware [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 991.594603] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b6dd2c-db4b-4506-99ca-7ee75e602b6c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.605136] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f5ef48-f2bd-4665-831a-0cdecd676ec5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.702457] env[61936]: DEBUG oslo_vmware.api [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253298, 'name': PowerOnVM_Task, 'duration_secs': 0.674727} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.702726] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.702926] env[61936]: INFO nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Took 15.17 seconds to spawn the instance on the hypervisor. [ 991.703119] env[61936]: DEBUG nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 991.703894] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5dbc42a-1159-4c0a-9a66-2b5a0313abc0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.895042] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2c50f939-f907-4111-aa92-7050457e3fbd tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.197s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.041381] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253299, 'name': Rename_Task, 'duration_secs': 0.14193} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.041381] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.041381] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c3b7ca9-a0de-4204-9613-1e59dc57bbf9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.051231] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 992.051231] env[61936]: value = "task-1253300" [ 992.051231] env[61936]: _type = "Task" [ 992.051231] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.059919] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253300, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.232107] env[61936]: INFO nova.compute.manager [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Took 33.66 seconds to build instance. [ 992.315622] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52de195b-1584-dbcb-9d9c-f9975ea22015/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 992.317680] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678f1f04-18b6-4d2c-b846-9bc7ac10e23b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.325106] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52de195b-1584-dbcb-9d9c-f9975ea22015/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 992.325492] env[61936]: ERROR oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52de195b-1584-dbcb-9d9c-f9975ea22015/disk-0.vmdk due to incomplete transfer. [ 992.325906] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7b2dd746-7acc-49e1-bae3-c1438d96b6ed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.333742] env[61936]: DEBUG oslo_vmware.rw_handles [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52de195b-1584-dbcb-9d9c-f9975ea22015/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 992.333950] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Uploaded image e8676683-128d-4455-aff8-ca32f97ce13e to the Glance image server {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 992.336585] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Destroying the VM {{(pid=61936) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 992.336848] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aa08e949-3a06-4163-83f3-fcf8ceff2aaf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.342448] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 992.342448] env[61936]: value = "task-1253301" [ 992.342448] env[61936]: _type = "Task" [ 992.342448] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.351018] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253301, 'name': Destroy_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.566382] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253300, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.659448] env[61936]: DEBUG nova.compute.manager [req-32d15521-1549-4a85-9bfa-2644b44cf8d3 req-233a4e9f-40c0-4369-94bc-3d8cbd3b61c5 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Received event network-vif-plugged-e21d453a-d68d-4497-9042-ebc1a0dbec83 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 992.659448] env[61936]: DEBUG oslo_concurrency.lockutils [req-32d15521-1549-4a85-9bfa-2644b44cf8d3 req-233a4e9f-40c0-4369-94bc-3d8cbd3b61c5 service nova] Acquiring lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.659663] env[61936]: DEBUG oslo_concurrency.lockutils [req-32d15521-1549-4a85-9bfa-2644b44cf8d3 req-233a4e9f-40c0-4369-94bc-3d8cbd3b61c5 service nova] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.659818] env[61936]: DEBUG oslo_concurrency.lockutils [req-32d15521-1549-4a85-9bfa-2644b44cf8d3 req-233a4e9f-40c0-4369-94bc-3d8cbd3b61c5 service nova] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.660025] env[61936]: DEBUG nova.compute.manager [req-32d15521-1549-4a85-9bfa-2644b44cf8d3 req-233a4e9f-40c0-4369-94bc-3d8cbd3b61c5 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] No waiting events found dispatching network-vif-plugged-e21d453a-d68d-4497-9042-ebc1a0dbec83 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 992.660237] env[61936]: WARNING nova.compute.manager [req-32d15521-1549-4a85-9bfa-2644b44cf8d3 req-233a4e9f-40c0-4369-94bc-3d8cbd3b61c5 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Received unexpected event network-vif-plugged-e21d453a-d68d-4497-9042-ebc1a0dbec83 for instance with vm_state building and task_state spawning. [ 992.715786] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6101fc65-9c76-4efc-90fc-bcb6bf4b2c0e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.726834] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e09483-4867-4e45-a06c-08afc5419989 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.731916] env[61936]: DEBUG oslo_concurrency.lockutils [None req-74c1e940-bb77-428e-a6a5-29622afb8796 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.174s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.762907] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781a6fc2-57c3-418b-9341-70685a0721d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.770902] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d25f71-f7a8-4117-adae-dc93ac05efbd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.790741] env[61936]: DEBUG nova.compute.provider_tree [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.816067] env[61936]: DEBUG nova.network.neutron [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Successfully updated port: e21d453a-d68d-4497-9042-ebc1a0dbec83 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.852275] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253301, 'name': Destroy_Task, 'duration_secs': 0.387064} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.852568] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Destroyed the VM [ 992.852827] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Deleting Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 992.853095] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-df0b5c9e-db38-4869-822b-ddeaee11d1bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.859446] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 992.859446] env[61936]: value = "task-1253302" [ 992.859446] env[61936]: _type = "Task" [ 992.859446] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.868338] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253302, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.063554] env[61936]: DEBUG oslo_vmware.api [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253300, 'name': PowerOnVM_Task, 'duration_secs': 0.523027} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.063936] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 993.064054] env[61936]: INFO nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Took 9.85 seconds to spawn the instance on the hypervisor. [ 993.064238] env[61936]: DEBUG nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 993.065169] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541bafa5-a3b2-4c1f-a92c-ca6803463f48 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.294813] env[61936]: DEBUG nova.scheduler.client.report [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 993.318622] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-90c2fd2c-937c-4b92-8bb1-6387652a03bb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.318798] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-90c2fd2c-937c-4b92-8bb1-6387652a03bb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.318948] env[61936]: DEBUG nova.network.neutron [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.370384] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253302, 'name': RemoveSnapshot_Task, 'duration_secs': 0.402295} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.370677] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Deleted Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 993.370979] env[61936]: DEBUG nova.compute.manager [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 993.371796] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba40515-6dd7-44af-bc31-ef9b04118ed9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.587691] env[61936]: INFO nova.compute.manager [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Took 31.15 seconds to build instance. [ 993.724090] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "62b33dd9-e523-4309-94e6-59671fcfbaea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.724381] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.724600] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "62b33dd9-e523-4309-94e6-59671fcfbaea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.724838] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.725056] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.728948] env[61936]: INFO nova.compute.manager [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Terminating instance [ 993.744125] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.744125] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.800392] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.800946] env[61936]: DEBUG nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 993.803404] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.787s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.804743] env[61936]: INFO nova.compute.claims [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.849803] env[61936]: DEBUG nova.network.neutron [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.886911] env[61936]: INFO nova.compute.manager [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Shelve offloading [ 994.004224] env[61936]: DEBUG nova.network.neutron [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Updating instance_info_cache with network_info: [{"id": "e21d453a-d68d-4497-9042-ebc1a0dbec83", "address": "fa:16:3e:bf:cd:76", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d453a-d6", "ovs_interfaceid": "e21d453a-d68d-4497-9042-ebc1a0dbec83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.090389] env[61936]: DEBUG oslo_concurrency.lockutils [None req-7142dd3a-5b25-4900-84d0-9af1f70d94cf tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.659s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.233422] env[61936]: DEBUG nova.compute.manager [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 994.233639] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.234622] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae330059-ebf0-4ddf-9bd2-66e629a86c13 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.243063] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.243343] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a54db483-d34d-4de8-bc1e-1611be8ef26c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.245493] env[61936]: DEBUG nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 994.250364] env[61936]: DEBUG oslo_vmware.api [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 994.250364] env[61936]: value = "task-1253303" [ 994.250364] env[61936]: _type = "Task" [ 994.250364] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.259132] env[61936]: DEBUG oslo_vmware.api [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.311031] env[61936]: DEBUG nova.compute.utils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 994.313408] env[61936]: DEBUG nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 994.313824] env[61936]: DEBUG nova.network.neutron [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 994.396545] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.396545] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84f0bb7b-6c38-4a4c-a167-c84e502a231f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.403647] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 994.403647] env[61936]: value = "task-1253304" [ 994.403647] env[61936]: _type = "Task" [ 994.403647] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.413251] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253304, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.466019] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.466494] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.493067] env[61936]: DEBUG nova.policy [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d326aaaa0904babb398a804980a518f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43b25cccc1b44eeabee94f0810b58745', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 994.510017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-90c2fd2c-937c-4b92-8bb1-6387652a03bb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.510017] env[61936]: DEBUG nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Instance network_info: |[{"id": "e21d453a-d68d-4497-9042-ebc1a0dbec83", "address": "fa:16:3e:bf:cd:76", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d453a-d6", "ovs_interfaceid": "e21d453a-d68d-4497-9042-ebc1a0dbec83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 994.510017] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:cd:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e21d453a-d68d-4497-9042-ebc1a0dbec83', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.516421] env[61936]: DEBUG oslo.service.loopingcall [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.517169] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.517546] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ce3c35b-fd32-43c7-a72a-e95577b50b09 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.540690] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.540690] env[61936]: value = "task-1253305" [ 994.540690] env[61936]: _type = "Task" [ 994.540690] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.547917] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253305, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.772271] env[61936]: DEBUG oslo_vmware.api [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253303, 'name': PowerOffVM_Task, 'duration_secs': 0.368282} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.772674] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 994.773129] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.773492] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7434189-d885-4927-abda-a1b103beda4c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.782622] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.814793] env[61936]: DEBUG nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 994.844998] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.845251] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.849384] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleting the datastore file [datastore1] 62b33dd9-e523-4309-94e6-59671fcfbaea {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.849384] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-910f8896-0137-4058-ac8e-9097694e86af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.854803] env[61936]: DEBUG oslo_vmware.api [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 994.854803] env[61936]: value = "task-1253307" [ 994.854803] env[61936]: _type = "Task" [ 994.854803] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.864782] env[61936]: DEBUG oslo_vmware.api [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.872546] env[61936]: DEBUG nova.compute.manager [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Received event network-changed-e21d453a-d68d-4497-9042-ebc1a0dbec83 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 994.872806] env[61936]: DEBUG nova.compute.manager [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Refreshing instance network info cache due to event network-changed-e21d453a-d68d-4497-9042-ebc1a0dbec83. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 994.873130] env[61936]: DEBUG oslo_concurrency.lockutils [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] Acquiring lock "refresh_cache-90c2fd2c-937c-4b92-8bb1-6387652a03bb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.874570] env[61936]: DEBUG oslo_concurrency.lockutils [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] Acquired lock "refresh_cache-90c2fd2c-937c-4b92-8bb1-6387652a03bb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.874570] env[61936]: DEBUG nova.network.neutron [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Refreshing network info cache for port e21d453a-d68d-4497-9042-ebc1a0dbec83 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.885567] env[61936]: DEBUG nova.compute.manager [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Received event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 994.885780] env[61936]: DEBUG nova.compute.manager [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing instance network info cache due to event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 994.886019] env[61936]: DEBUG oslo_concurrency.lockutils [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] Acquiring lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.886217] env[61936]: DEBUG oslo_concurrency.lockutils [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] Acquired lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.886384] env[61936]: DEBUG nova.network.neutron [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.916762] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 994.917051] env[61936]: DEBUG nova.compute.manager [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 994.917989] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a3c446-e241-4160-9338-fad6326ac5f0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.928891] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.929179] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.929462] env[61936]: DEBUG nova.network.neutron [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 994.968872] env[61936]: DEBUG nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 995.000854] env[61936]: DEBUG nova.network.neutron [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Successfully created port: 4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 995.052382] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253305, 'name': CreateVM_Task, 'duration_secs': 0.355213} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.052561] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.053267] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.053437] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.053760] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.054038] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01505f80-6150-4cc9-883a-8a1339e0c0d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.062868] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 995.062868] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b1517f-2791-f964-5b48-3a5b3e241ee5" [ 995.062868] env[61936]: _type = "Task" [ 995.062868] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.073946] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b1517f-2791-f964-5b48-3a5b3e241ee5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.099301] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222e2ac4-b1bd-4d0a-9517-2931d7c83ee8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.106915] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44496fd0-497d-4029-b434-796b5eb7be85 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.137650] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a039488c-6b15-49c1-8e5b-482c92fb55ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.145830] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb97acb-9106-47f6-9d74-6f15b0334f84 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.160028] env[61936]: DEBUG nova.compute.provider_tree [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.365307] env[61936]: DEBUG oslo_vmware.api [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218182} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.365726] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.365726] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.365893] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.366074] env[61936]: INFO nova.compute.manager [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Took 1.13 seconds to destroy the instance on the hypervisor. [ 995.366312] env[61936]: DEBUG oslo.service.loopingcall [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.366511] env[61936]: DEBUG nova.compute.manager [-] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 995.366601] env[61936]: DEBUG nova.network.neutron [-] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.492305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.584152] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52b1517f-2791-f964-5b48-3a5b3e241ee5, 'name': SearchDatastore_Task, 'duration_secs': 0.011005} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.585574] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.586067] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.586668] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.589017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.589017] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.589017] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23171cfa-c059-41ad-bad1-707ce46bc84e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.603017] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.603017] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.603017] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01eb9d22-8744-4207-a5f4-a86242fab689 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.612962] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 995.612962] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e09575-7a9f-cf7e-ff80-f6b076353146" [ 995.612962] env[61936]: _type = "Task" [ 995.612962] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.621126] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e09575-7a9f-cf7e-ff80-f6b076353146, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.661966] env[61936]: DEBUG nova.scheduler.client.report [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.726443] env[61936]: DEBUG nova.network.neutron [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Updated VIF entry in instance network info cache for port e21d453a-d68d-4497-9042-ebc1a0dbec83. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.726811] env[61936]: DEBUG nova.network.neutron [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Updating instance_info_cache with network_info: [{"id": "e21d453a-d68d-4497-9042-ebc1a0dbec83", "address": "fa:16:3e:bf:cd:76", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21d453a-d6", "ovs_interfaceid": "e21d453a-d68d-4497-9042-ebc1a0dbec83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.763294] env[61936]: DEBUG nova.network.neutron [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Updating instance_info_cache with network_info: [{"id": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "address": "fa:16:3e:ee:f9:9e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82ce2f0e-a7", "ovs_interfaceid": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.828191] env[61936]: DEBUG nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 995.861685] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 995.861940] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.862298] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 995.862298] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.862480] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 995.862582] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 995.862790] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 995.862948] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 995.863128] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 995.863294] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 995.863466] env[61936]: DEBUG nova.virt.hardware [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 995.864351] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a3b62c-4c89-44dc-aa89-797db0d3366b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.872558] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e176c5a5-4571-4f76-9168-8674d1108355 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.932684] env[61936]: DEBUG nova.network.neutron [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updated VIF entry in instance network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.933156] env[61936]: DEBUG nova.network.neutron [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.123293] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e09575-7a9f-cf7e-ff80-f6b076353146, 'name': SearchDatastore_Task, 'duration_secs': 0.026916} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.124136] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3a2dc13-1a31-40dd-a600-f3830516f8e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.128974] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 996.128974] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ef2bcb-e1db-1fe7-71ce-942aed2f9e14" [ 996.128974] env[61936]: _type = "Task" [ 996.128974] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.136153] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ef2bcb-e1db-1fe7-71ce-942aed2f9e14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.167118] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.167521] env[61936]: DEBUG nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 996.170323] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.070s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.170518] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.172582] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.418s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.172793] env[61936]: DEBUG nova.objects.instance [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lazy-loading 'pci_requests' on Instance uuid 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.193664] env[61936]: INFO nova.scheduler.client.report [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted allocations for instance 9bd97f84-7c21-44ad-9ed0-d7e4097e264e [ 996.229561] env[61936]: DEBUG oslo_concurrency.lockutils [req-78d9f3d1-c9a0-41ef-be8d-fa96c449f0a2 req-567535de-029d-4708-88bd-9fc927269f32 service nova] Releasing lock "refresh_cache-90c2fd2c-937c-4b92-8bb1-6387652a03bb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.264709] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.393307] env[61936]: DEBUG nova.network.neutron [-] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.435681] env[61936]: DEBUG oslo_concurrency.lockutils [req-519f76e9-1083-4065-b115-e3fca2a51d47 req-ea679816-35b9-4eec-bbf8-342c5b0b815d service nova] Releasing lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.444140] env[61936]: DEBUG nova.compute.manager [req-0fd7ecd7-58e7-4c24-9a63-bbfc16e2c769 req-dcba04e7-28d2-486f-a202-ad3b9046eb4d service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Received event network-vif-plugged-4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 996.444354] env[61936]: DEBUG oslo_concurrency.lockutils [req-0fd7ecd7-58e7-4c24-9a63-bbfc16e2c769 req-dcba04e7-28d2-486f-a202-ad3b9046eb4d service nova] Acquiring lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.444554] env[61936]: DEBUG oslo_concurrency.lockutils [req-0fd7ecd7-58e7-4c24-9a63-bbfc16e2c769 req-dcba04e7-28d2-486f-a202-ad3b9046eb4d service nova] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.444729] env[61936]: DEBUG oslo_concurrency.lockutils [req-0fd7ecd7-58e7-4c24-9a63-bbfc16e2c769 req-dcba04e7-28d2-486f-a202-ad3b9046eb4d service nova] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.444891] env[61936]: DEBUG nova.compute.manager [req-0fd7ecd7-58e7-4c24-9a63-bbfc16e2c769 req-dcba04e7-28d2-486f-a202-ad3b9046eb4d service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] No waiting events found dispatching network-vif-plugged-4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 996.445074] env[61936]: WARNING nova.compute.manager [req-0fd7ecd7-58e7-4c24-9a63-bbfc16e2c769 req-dcba04e7-28d2-486f-a202-ad3b9046eb4d service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Received unexpected event network-vif-plugged-4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b for instance with vm_state building and task_state spawning. [ 996.640690] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ef2bcb-e1db-1fe7-71ce-942aed2f9e14, 'name': SearchDatastore_Task, 'duration_secs': 0.044642} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.640955] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.641236] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 90c2fd2c-937c-4b92-8bb1-6387652a03bb/90c2fd2c-937c-4b92-8bb1-6387652a03bb.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.641693] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5a1acb39-cf13-49f0-8422-a86e31938eb9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.648900] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 996.648900] env[61936]: value = "task-1253308" [ 996.648900] env[61936]: _type = "Task" [ 996.648900] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.657722] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.669024] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.669800] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2f70e8-e957-4d03-85a0-ab9e35665e81 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.675722] env[61936]: DEBUG nova.compute.utils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.678181] env[61936]: DEBUG nova.objects.instance [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lazy-loading 'numa_topology' on Instance uuid 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.681391] env[61936]: DEBUG nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 996.681556] env[61936]: DEBUG nova.network.neutron [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.683253] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.683648] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ac282e5-49e3-4214-b945-bedf4e1e2ed8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.700668] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0e7bf441-d037-48a5-bce5-63a2371d2dff tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "9bd97f84-7c21-44ad-9ed0-d7e4097e264e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.360s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.725469] env[61936]: DEBUG nova.policy [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '288fd1d0d8e341999481fc62fe14195e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f1b64c32564b9cbb68e3799b456641', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.792341] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.792744] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.793078] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleting the datastore file [datastore2] dc6523b7-c15c-469f-83a5-5c5dc5f2d230 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.793511] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e57e6a3-db76-4850-b02d-b2307f3b2e36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.801742] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 996.801742] env[61936]: value = "task-1253310" [ 996.801742] env[61936]: _type = "Task" [ 996.801742] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.811552] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253310, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.894831] env[61936]: INFO nova.compute.manager [-] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Took 1.53 seconds to deallocate network for instance. [ 996.913882] env[61936]: DEBUG nova.compute.manager [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Received event network-vif-deleted-18af91fe-34f3-43a1-9f99-73143ee13d3f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 996.913882] env[61936]: DEBUG nova.compute.manager [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Received event network-vif-unplugged-82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 996.913882] env[61936]: DEBUG oslo_concurrency.lockutils [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] Acquiring lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.913882] env[61936]: DEBUG oslo_concurrency.lockutils [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.913882] env[61936]: DEBUG oslo_concurrency.lockutils [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.914162] env[61936]: DEBUG nova.compute.manager [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] No waiting events found dispatching network-vif-unplugged-82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 996.916079] env[61936]: WARNING nova.compute.manager [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Received unexpected event network-vif-unplugged-82ce2f0e-a794-4a8a-a494-0b9deb336108 for instance with vm_state shelved and task_state shelving_offloading. [ 996.916079] env[61936]: DEBUG nova.compute.manager [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Received event network-changed-82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 996.916079] env[61936]: DEBUG nova.compute.manager [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Refreshing instance network info cache due to event network-changed-82ce2f0e-a794-4a8a-a494-0b9deb336108. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 996.916079] env[61936]: DEBUG oslo_concurrency.lockutils [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] Acquiring lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.916079] env[61936]: DEBUG oslo_concurrency.lockutils [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] Acquired lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.916079] env[61936]: DEBUG nova.network.neutron [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Refreshing network info cache for port 82ce2f0e-a794-4a8a-a494-0b9deb336108 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.130789] env[61936]: DEBUG nova.network.neutron [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Successfully updated port: 4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.167294] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253308, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449445} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.167725] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 90c2fd2c-937c-4b92-8bb1-6387652a03bb/90c2fd2c-937c-4b92-8bb1-6387652a03bb.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.168168] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.168550] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc80058f-2822-4c61-8a16-c376057322cd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.178218] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 997.178218] env[61936]: value = "task-1253311" [ 997.178218] env[61936]: _type = "Task" [ 997.178218] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.182252] env[61936]: DEBUG nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 997.187964] env[61936]: INFO nova.compute.claims [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.190506] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253311, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.233205] env[61936]: DEBUG nova.network.neutron [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Successfully created port: 7775677e-cf35-458b-8c94-da7dd8dc652a {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.311731] env[61936]: DEBUG oslo_vmware.api [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253310, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314621} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.311984] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.312190] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.312367] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.334991] env[61936]: INFO nova.scheduler.client.report [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted allocations for instance dc6523b7-c15c-469f-83a5-5c5dc5f2d230 [ 997.402475] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.633247] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "refresh_cache-8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.633390] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquired lock "refresh_cache-8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.633539] env[61936]: DEBUG nova.network.neutron [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.684368] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253311, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061572} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.684637] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.685406] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d17865d-cd76-41de-8bbc-974dba633594 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.713253] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 90c2fd2c-937c-4b92-8bb1-6387652a03bb/90c2fd2c-937c-4b92-8bb1-6387652a03bb.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.714592] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f68cd800-48fb-4a6d-92cc-1213c90d9125 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.737368] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 997.737368] env[61936]: value = "task-1253312" [ 997.737368] env[61936]: _type = "Task" [ 997.737368] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.746053] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253312, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.773623] env[61936]: DEBUG nova.network.neutron [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Updated VIF entry in instance network info cache for port 82ce2f0e-a794-4a8a-a494-0b9deb336108. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 997.774105] env[61936]: DEBUG nova.network.neutron [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Updating instance_info_cache with network_info: [{"id": "82ce2f0e-a794-4a8a-a494-0b9deb336108", "address": "fa:16:3e:ee:f9:9e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": null, "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap82ce2f0e-a7", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.839363] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.167244] env[61936]: DEBUG nova.network.neutron [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 998.215271] env[61936]: DEBUG nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 998.251983] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253312, 'name': ReconfigVM_Task, 'duration_secs': 0.302016} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.254082] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 998.254328] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.254486] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 998.254669] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.254816] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 998.254963] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 998.255189] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 998.255348] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 998.255513] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 998.255674] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 998.255846] env[61936]: DEBUG nova.virt.hardware [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 998.256166] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 90c2fd2c-937c-4b92-8bb1-6387652a03bb/90c2fd2c-937c-4b92-8bb1-6387652a03bb.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.257363] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d1394d-d244-4ea9-bd29-0cee3d3b7d15 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.259853] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48541115-0831-420f-bf87-e88691dd9eed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.270556] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29cbd0c-df75-437c-bb3c-4d0ab15055a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.274801] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 998.274801] env[61936]: value = "task-1253313" [ 998.274801] env[61936]: _type = "Task" [ 998.274801] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.278301] env[61936]: DEBUG oslo_concurrency.lockutils [req-26304459-993c-4f2d-b5d9-477f12a4bd99 req-a8a19037-9f22-4f6d-b469-5169aa518ffb service nova] Releasing lock "refresh_cache-dc6523b7-c15c-469f-83a5-5c5dc5f2d230" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.301394] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253313, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.406486] env[61936]: DEBUG nova.network.neutron [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Updating instance_info_cache with network_info: [{"id": "4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b", "address": "fa:16:3e:1a:6a:f9", "network": {"id": "4adecfba-ff5c-4c71-856e-c386acc74743", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-658655058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43b25cccc1b44eeabee94f0810b58745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ecf3c38-78", "ovs_interfaceid": "4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.440157] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa8aac8-8db5-4b92-b738-bf4ad2cae21e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.447872] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b964c9-60d9-46e8-a83f-0695462cf3fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.481076] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28beabb-3724-4c9e-ac7a-3adfcfefb571 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.491889] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6175be-ac34-456b-8cfe-983482870a94 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.506325] env[61936]: DEBUG nova.compute.provider_tree [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.787911] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253313, 'name': Rename_Task, 'duration_secs': 0.134246} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.790557] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.790557] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9cf49f1f-ba01-4161-9f9c-79ba5e8fbf27 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.797163] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 998.797163] env[61936]: value = "task-1253314" [ 998.797163] env[61936]: _type = "Task" [ 998.797163] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.804207] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253314, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.913017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Releasing lock "refresh_cache-8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.913017] env[61936]: DEBUG nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Instance network_info: |[{"id": "4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b", "address": "fa:16:3e:1a:6a:f9", "network": {"id": "4adecfba-ff5c-4c71-856e-c386acc74743", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-658655058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43b25cccc1b44eeabee94f0810b58745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ecf3c38-78", "ovs_interfaceid": "4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 998.913017] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:6a:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cd38fdec-d092-4a84-ab41-685f6dbb4f29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 998.920024] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Creating folder: Project (43b25cccc1b44eeabee94f0810b58745). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 998.921261] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b3838ee-a08b-40f0-9a93-b124034f88ba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.932734] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Created folder: Project (43b25cccc1b44eeabee94f0810b58745) in parent group-v269874. [ 998.933122] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Creating folder: Instances. Parent ref: group-v270008. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 998.935091] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04a33a25-e2b6-44ee-9470-f5d72d822f54 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.943786] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Created folder: Instances in parent group-v270008. [ 998.943786] env[61936]: DEBUG oslo.service.loopingcall [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.945406] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 998.946898] env[61936]: DEBUG nova.compute.manager [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Received event network-changed-4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 998.948033] env[61936]: DEBUG nova.compute.manager [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Refreshing instance network info cache due to event network-changed-4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 998.948033] env[61936]: DEBUG oslo_concurrency.lockutils [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] Acquiring lock "refresh_cache-8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.948033] env[61936]: DEBUG oslo_concurrency.lockutils [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] Acquired lock "refresh_cache-8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.948033] env[61936]: DEBUG nova.network.neutron [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Refreshing network info cache for port 4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 998.949246] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d20074f9-c819-4022-be7c-364bd83b5e58 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.969104] env[61936]: DEBUG nova.network.neutron [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Successfully updated port: 7775677e-cf35-458b-8c94-da7dd8dc652a {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.976107] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 998.976107] env[61936]: value = "task-1253317" [ 998.976107] env[61936]: _type = "Task" [ 998.976107] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.985308] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253317, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.988088] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.008333] env[61936]: DEBUG nova.scheduler.client.report [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 999.225952] env[61936]: DEBUG nova.network.neutron [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Updated VIF entry in instance network info cache for port 4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 999.226345] env[61936]: DEBUG nova.network.neutron [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Updating instance_info_cache with network_info: [{"id": "4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b", "address": "fa:16:3e:1a:6a:f9", "network": {"id": "4adecfba-ff5c-4c71-856e-c386acc74743", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-658655058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43b25cccc1b44eeabee94f0810b58745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd38fdec-d092-4a84-ab41-685f6dbb4f29", "external-id": "nsx-vlan-transportzone-622", "segmentation_id": 622, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ecf3c38-78", "ovs_interfaceid": "4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.309290] env[61936]: DEBUG oslo_vmware.api [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253314, 'name': PowerOnVM_Task, 'duration_secs': 0.453096} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.309704] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.310028] env[61936]: INFO nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Took 7.75 seconds to spawn the instance on the hypervisor. [ 999.310347] env[61936]: DEBUG nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 999.311475] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289ec200-732a-4c29-b38c-0ee798173e7d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.471436] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.471742] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.471742] env[61936]: DEBUG nova.network.neutron [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.485310] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253317, 'name': CreateVM_Task, 'duration_secs': 0.47681} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.485481] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 999.486146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.486315] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.486637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 999.486887] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e0fc63c-6c16-4562-ac88-43e249bc6822 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.492331] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 999.492331] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e8ea8c-ae1b-f033-1ddd-c40c241435b3" [ 999.492331] env[61936]: _type = "Task" [ 999.492331] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.500321] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e8ea8c-ae1b-f033-1ddd-c40c241435b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.512441] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.340s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.514407] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.732s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.515959] env[61936]: INFO nova.compute.claims [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.547073] env[61936]: INFO nova.network.neutron [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating port 1cdd9a15-01e7-40cf-80fc-24bea0b642d3 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 999.729773] env[61936]: DEBUG oslo_concurrency.lockutils [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] Releasing lock "refresh_cache-8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.730119] env[61936]: DEBUG nova.compute.manager [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Received event network-vif-plugged-7775677e-cf35-458b-8c94-da7dd8dc652a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 999.730366] env[61936]: DEBUG oslo_concurrency.lockutils [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] Acquiring lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.730577] env[61936]: DEBUG oslo_concurrency.lockutils [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.730740] env[61936]: DEBUG oslo_concurrency.lockutils [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.730953] env[61936]: DEBUG nova.compute.manager [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] No waiting events found dispatching network-vif-plugged-7775677e-cf35-458b-8c94-da7dd8dc652a {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 999.731090] env[61936]: WARNING nova.compute.manager [req-7e749f90-6617-4464-a8bd-53c0c86c115d req-4ff4794e-de72-4de9-968d-2ecdc2a3ef13 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Received unexpected event network-vif-plugged-7775677e-cf35-458b-8c94-da7dd8dc652a for instance with vm_state building and task_state spawning. [ 999.836108] env[61936]: INFO nova.compute.manager [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Took 23.86 seconds to build instance. [ 1000.003423] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e8ea8c-ae1b-f033-1ddd-c40c241435b3, 'name': SearchDatastore_Task, 'duration_secs': 0.042011} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.003720] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.003954] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.004199] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.004349] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.004528] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.004799] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c438c0d-88ba-474c-a06a-bb50c83def62 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.012805] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.012985] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.013803] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ea4851b-edae-4662-8f8f-9ea9e5b46254 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.018787] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1000.018787] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522670e0-12b3-d150-d971-db2cc9fa799a" [ 1000.018787] env[61936]: _type = "Task" [ 1000.018787] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.029150] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522670e0-12b3-d150-d971-db2cc9fa799a, 'name': SearchDatastore_Task, 'duration_secs': 0.007658} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.029984] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-648a09ad-b0a3-4ee3-8f34-ddd8d17bc87d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.035415] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1000.035415] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c48a0e-4035-9a68-f7ac-e09107695141" [ 1000.035415] env[61936]: _type = "Task" [ 1000.035415] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.043276] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c48a0e-4035-9a68-f7ac-e09107695141, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.046376] env[61936]: DEBUG nova.network.neutron [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.337517] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3125181f-3224-46a5-b966-789862e786ab tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.373s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.518138] env[61936]: DEBUG nova.network.neutron [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance_info_cache with network_info: [{"id": "7775677e-cf35-458b-8c94-da7dd8dc652a", "address": "fa:16:3e:f7:93:44", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7775677e-cf", "ovs_interfaceid": "7775677e-cf35-458b-8c94-da7dd8dc652a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.547565] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c48a0e-4035-9a68-f7ac-e09107695141, 'name': SearchDatastore_Task, 'duration_secs': 0.013728} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.550462] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.550721] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6/8b6696e5-ef1b-4825-903f-de4dfb9bf2a6.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.551545] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73087dd7-ae83-4cfa-a7c4-cc2bb6212199 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.558857] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1000.558857] env[61936]: value = "task-1253318" [ 1000.558857] env[61936]: _type = "Task" [ 1000.558857] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.569859] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253318, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.596715] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.596998] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.597232] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.597395] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.597561] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.601670] env[61936]: INFO nova.compute.manager [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Terminating instance [ 1000.726045] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f526d38-1385-48b1-b061-e5561ae7c67d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.735135] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4450fb29-aa0a-417d-aed8-c3c252e73e26 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.767388] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0433b105-d8ac-4aab-bbca-eefff110154a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.775173] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b41ee1-2d38-4a1d-94b4-5f5a476415d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.790274] env[61936]: DEBUG nova.compute.provider_tree [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.972177] env[61936]: DEBUG nova.compute.manager [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Received event network-changed-7775677e-cf35-458b-8c94-da7dd8dc652a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1000.972566] env[61936]: DEBUG nova.compute.manager [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Refreshing instance network info cache due to event network-changed-7775677e-cf35-458b-8c94-da7dd8dc652a. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1000.972820] env[61936]: DEBUG oslo_concurrency.lockutils [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] Acquiring lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.020547] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.020882] env[61936]: DEBUG nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Instance network_info: |[{"id": "7775677e-cf35-458b-8c94-da7dd8dc652a", "address": "fa:16:3e:f7:93:44", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7775677e-cf", "ovs_interfaceid": "7775677e-cf35-458b-8c94-da7dd8dc652a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1001.021239] env[61936]: DEBUG oslo_concurrency.lockutils [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] Acquired lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.021420] env[61936]: DEBUG nova.network.neutron [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Refreshing network info cache for port 7775677e-cf35-458b-8c94-da7dd8dc652a {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1001.022670] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:93:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b94712a6-b777-47dd-bc06-f9acfce2d936', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7775677e-cf35-458b-8c94-da7dd8dc652a', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.031513] env[61936]: DEBUG oslo.service.loopingcall [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.034401] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1001.034859] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96651ff8-376d-4f49-b8e1-085200a9c701 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.056639] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.056639] env[61936]: value = "task-1253319" [ 1001.056639] env[61936]: _type = "Task" [ 1001.056639] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.067481] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253319, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.070444] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253318, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440819} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.070682] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6/8b6696e5-ef1b-4825-903f-de4dfb9bf2a6.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1001.070886] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.071166] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9070dda8-7fdc-4590-9325-1eaf8c904115 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.078538] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1001.078538] env[61936]: value = "task-1253320" [ 1001.078538] env[61936]: _type = "Task" [ 1001.078538] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.089944] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253320, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.106098] env[61936]: DEBUG nova.compute.manager [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1001.106098] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.107398] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b764ad4-2040-46f7-bb68-ee5d1852cd37 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.117611] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.117917] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dd1052d-b2c6-40a7-8bcb-f1da2976d46d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.124269] env[61936]: DEBUG oslo_vmware.api [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1001.124269] env[61936]: value = "task-1253321" [ 1001.124269] env[61936]: _type = "Task" [ 1001.124269] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.132751] env[61936]: DEBUG oslo_vmware.api [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253321, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.293875] env[61936]: DEBUG nova.scheduler.client.report [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1001.318537] env[61936]: DEBUG nova.network.neutron [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updated VIF entry in instance network info cache for port 7775677e-cf35-458b-8c94-da7dd8dc652a. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1001.319019] env[61936]: DEBUG nova.network.neutron [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance_info_cache with network_info: [{"id": "7775677e-cf35-458b-8c94-da7dd8dc652a", "address": "fa:16:3e:f7:93:44", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7775677e-cf", "ovs_interfaceid": "7775677e-cf35-458b-8c94-da7dd8dc652a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.568178] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253319, 'name': CreateVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.590033] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253320, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06841} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.590752] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.591578] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d33f40-5538-4b98-8f81-10c43e7b72ed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.614984] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6/8b6696e5-ef1b-4825-903f-de4dfb9bf2a6.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.615342] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54f32467-367d-4896-9d5a-10cfa29b124a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.639187] env[61936]: DEBUG oslo_vmware.api [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253321, 'name': PowerOffVM_Task, 'duration_secs': 0.196897} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.640440] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.640809] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.641136] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1001.641136] env[61936]: value = "task-1253322" [ 1001.641136] env[61936]: _type = "Task" [ 1001.641136] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.641336] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-138f006e-2df2-4977-bf9f-621e64195554 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.651804] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253322, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.756735] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.756735] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.756913] env[61936]: DEBUG nova.network.neutron [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.799055] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.284s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.799717] env[61936]: DEBUG nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1001.803022] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.311s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.804603] env[61936]: INFO nova.compute.claims [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.811188] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.811439] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.811673] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleting the datastore file [datastore1] 90c2fd2c-937c-4b92-8bb1-6387652a03bb {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.811969] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9188bcce-a70e-4dfa-99e5-55b8a904c452 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.818528] env[61936]: DEBUG oslo_vmware.api [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1001.818528] env[61936]: value = "task-1253324" [ 1001.818528] env[61936]: _type = "Task" [ 1001.818528] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.822398] env[61936]: DEBUG oslo_concurrency.lockutils [req-cd48d4aa-8832-485b-99bf-4868419e1686 req-a8f1f982-65a1-4af3-8d77-a4bab9ebcbb7 service nova] Releasing lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.828252] env[61936]: DEBUG oslo_vmware.api [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.067732] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253319, 'name': CreateVM_Task, 'duration_secs': 0.753879} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.067931] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1002.068646] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.068862] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.069223] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.069550] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b5691e9-823a-44c5-8e70-f63e33d0b85b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.074397] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1002.074397] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522897bf-6ac2-4977-1b2e-53284797f755" [ 1002.074397] env[61936]: _type = "Task" [ 1002.074397] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.082780] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522897bf-6ac2-4977-1b2e-53284797f755, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.159054] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253322, 'name': ReconfigVM_Task, 'duration_secs': 0.307435} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.159711] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6/8b6696e5-ef1b-4825-903f-de4dfb9bf2a6.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.159945] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e43bee3-cd6a-4954-935d-bcc554ad339a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.167235] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1002.167235] env[61936]: value = "task-1253325" [ 1002.167235] env[61936]: _type = "Task" [ 1002.167235] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.175385] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253325, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.313367] env[61936]: DEBUG nova.compute.utils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1002.314722] env[61936]: DEBUG nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1002.314883] env[61936]: DEBUG nova.network.neutron [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1002.329035] env[61936]: DEBUG oslo_vmware.api [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.416204} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.329284] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.329467] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.329639] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.329849] env[61936]: INFO nova.compute.manager [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Took 1.22 seconds to destroy the instance on the hypervisor. [ 1002.330144] env[61936]: DEBUG oslo.service.loopingcall [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.330337] env[61936]: DEBUG nova.compute.manager [-] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1002.330436] env[61936]: DEBUG nova.network.neutron [-] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1002.423432] env[61936]: DEBUG nova.policy [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f378ef8b8b5f4ae59c0d71dd1661bb59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa564b684410493fa0028fd345048e02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1002.586136] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522897bf-6ac2-4977-1b2e-53284797f755, 'name': SearchDatastore_Task, 'duration_secs': 0.051392} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.586593] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.586840] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.587147] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.587322] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.587507] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.587777] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-112760c1-cb92-4470-83e3-025a637c2ad3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.598399] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.598576] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1002.599332] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8563de2a-b3be-45b4-b06b-ac24950e35f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.604302] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1002.604302] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52738e17-73e2-9373-f1cb-6def86d4e308" [ 1002.604302] env[61936]: _type = "Task" [ 1002.604302] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.611686] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52738e17-73e2-9373-f1cb-6def86d4e308, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.681137] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253325, 'name': Rename_Task, 'duration_secs': 0.145572} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.685422] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1002.685787] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20e551a4-f32b-4dec-abeb-086e2e804bb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.694216] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1002.694216] env[61936]: value = "task-1253326" [ 1002.694216] env[61936]: _type = "Task" [ 1002.694216] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.705763] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253326, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.775634] env[61936]: DEBUG nova.network.neutron [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.819430] env[61936]: DEBUG nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1003.012347] env[61936]: DEBUG nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received event network-vif-plugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1003.012347] env[61936]: DEBUG oslo_concurrency.lockutils [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.012501] env[61936]: DEBUG oslo_concurrency.lockutils [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.012665] env[61936]: DEBUG oslo_concurrency.lockutils [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.012847] env[61936]: DEBUG nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] No waiting events found dispatching network-vif-plugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1003.013036] env[61936]: WARNING nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received unexpected event network-vif-plugged-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 for instance with vm_state shelved_offloaded and task_state spawning. [ 1003.013217] env[61936]: DEBUG nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received event network-changed-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1003.013377] env[61936]: DEBUG nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Refreshing instance network info cache due to event network-changed-1cdd9a15-01e7-40cf-80fc-24bea0b642d3. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1003.013616] env[61936]: DEBUG oslo_concurrency.lockutils [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.058623] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930a1255-4c32-4ab4-a8a3-6fd5cf7357b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.061900] env[61936]: DEBUG nova.network.neutron [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Successfully created port: 855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1003.068701] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2aa361-67fd-4c11-8ecc-8d6e92c4e704 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.101809] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718c3cfb-2dfa-4555-9c8e-2c8bbd493bf2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.115899] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930af5d4-4739-4d1c-bea3-8f04d2d92b01 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.119771] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52738e17-73e2-9373-f1cb-6def86d4e308, 'name': SearchDatastore_Task, 'duration_secs': 0.007485} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.120913] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25b1880f-6b73-472f-a8d6-f5a32315db8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.131206] env[61936]: DEBUG nova.compute.provider_tree [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.135820] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1003.135820] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520e501e-4db5-65f2-4599-a2f64ae82e47" [ 1003.135820] env[61936]: _type = "Task" [ 1003.135820] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.143696] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520e501e-4db5-65f2-4599-a2f64ae82e47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.206277] env[61936]: DEBUG oslo_vmware.api [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253326, 'name': PowerOnVM_Task, 'duration_secs': 0.450077} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.206277] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.206277] env[61936]: INFO nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Took 7.38 seconds to spawn the instance on the hypervisor. [ 1003.206277] env[61936]: DEBUG nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1003.206277] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038770f0-8735-49e4-ab6f-accbada85afd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.279059] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.281386] env[61936]: DEBUG oslo_concurrency.lockutils [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.281587] env[61936]: DEBUG nova.network.neutron [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Refreshing network info cache for port 1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1003.312622] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='79b3d157b07e2dbaa1e7bfbe0d07a6f0',container_format='bare',created_at=2024-10-10T16:50:15Z,direct_url=,disk_format='vmdk',id=2f04c09d-c433-442e-b891-3fa38ecc215e,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-752027649-shelved',owner='c334e9711bc64304ab017eb6ab3e2d23',properties=ImageMetaProps,protected=,size=31666688,status='active',tags=,updated_at=2024-10-10T16:50:29Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1003.312879] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.313047] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1003.313239] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.313389] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1003.313538] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1003.313744] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1003.313904] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1003.314090] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1003.314255] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1003.314423] env[61936]: DEBUG nova.virt.hardware [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1003.315930] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc60cee-7e00-4cfd-bb39-9e2623994ab7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.329770] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abca347a-36e1-4b1c-a223-241826643bf8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.343696] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:fd:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a44713-0af1-486e-bc0d-00e03a769fa4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1cdd9a15-01e7-40cf-80fc-24bea0b642d3', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1003.351386] env[61936]: DEBUG oslo.service.loopingcall [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.351643] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1003.351878] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83f46516-5b05-4db5-83a1-b8b455e55fa3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.371688] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1003.371688] env[61936]: value = "task-1253327" [ 1003.371688] env[61936]: _type = "Task" [ 1003.371688] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.379152] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253327, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.428716] env[61936]: DEBUG nova.network.neutron [-] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.635418] env[61936]: DEBUG nova.scheduler.client.report [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1003.648329] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]520e501e-4db5-65f2-4599-a2f64ae82e47, 'name': SearchDatastore_Task, 'duration_secs': 0.009159} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.648716] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.649080] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 755c5fea-e3d2-4cb3-b717-5f862401d7e1/755c5fea-e3d2-4cb3-b717-5f862401d7e1.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1003.649404] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b36be0e1-9a69-42a6-bf5f-4e802d62d197 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.656734] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1003.656734] env[61936]: value = "task-1253328" [ 1003.656734] env[61936]: _type = "Task" [ 1003.656734] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.666224] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253328, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.722591] env[61936]: INFO nova.compute.manager [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Took 24.24 seconds to build instance. [ 1003.830475] env[61936]: DEBUG nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1003.856167] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1003.856668] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.856879] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1003.857115] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.857329] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1003.857495] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1003.857808] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1003.857933] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1003.858118] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1003.858314] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1003.858492] env[61936]: DEBUG nova.virt.hardware [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1003.859578] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b423a263-a001-470d-9ed3-ed5ab1f7f182 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.869428] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3624c9-9e92-4789-af54-ff6151d366de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.882789] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253327, 'name': CreateVM_Task, 'duration_secs': 0.302554} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.898072] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.905641] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.905829] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.906248] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.906552] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c412d4d5-0cbc-492a-9121-ddb6909e76b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.912105] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1003.912105] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52246612-bcb8-4ff2-a7dc-f289672e1d30" [ 1003.912105] env[61936]: _type = "Task" [ 1003.912105] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.921666] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52246612-bcb8-4ff2-a7dc-f289672e1d30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.931499] env[61936]: INFO nova.compute.manager [-] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Took 1.60 seconds to deallocate network for instance. [ 1004.143984] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.144341] env[61936]: DEBUG nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1004.147136] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.745s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.147305] env[61936]: DEBUG nova.objects.instance [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lazy-loading 'resources' on Instance uuid 62b33dd9-e523-4309-94e6-59671fcfbaea {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.151013] env[61936]: DEBUG nova.network.neutron [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updated VIF entry in instance network info cache for port 1cdd9a15-01e7-40cf-80fc-24bea0b642d3. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1004.151394] env[61936]: DEBUG nova.network.neutron [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.168086] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253328, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462263} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.168086] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 755c5fea-e3d2-4cb3-b717-5f862401d7e1/755c5fea-e3d2-4cb3-b717-5f862401d7e1.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.168086] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.169454] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51b4efa0-5cfd-4c79-b8c8-7fa2fe6513f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.176034] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1004.176034] env[61936]: value = "task-1253329" [ 1004.176034] env[61936]: _type = "Task" [ 1004.176034] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.184476] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253329, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.224112] env[61936]: DEBUG oslo_concurrency.lockutils [None req-c64784b3-b6b9-4ffe-b801-4b200776fe0e tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.754s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.298032] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.298355] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.298751] env[61936]: INFO nova.compute.manager [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Shelving [ 1004.423767] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.423767] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Processing image 2f04c09d-c433-442e-b891-3fa38ecc215e {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.424161] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e/2f04c09d-c433-442e-b891-3fa38ecc215e.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.424161] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e/2f04c09d-c433-442e-b891-3fa38ecc215e.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.424363] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.424495] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd2eaf79-d693-432d-b1bb-95f5a605ee2a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.433092] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.433308] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1004.433968] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6c6333d-7d5a-4f75-9a7b-3b54405b6e8a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.439195] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1004.439195] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cf356a-020b-f705-46a7-0a406d2928b8" [ 1004.439195] env[61936]: _type = "Task" [ 1004.439195] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.440121] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.447590] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cf356a-020b-f705-46a7-0a406d2928b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.650513] env[61936]: DEBUG nova.compute.utils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1004.655528] env[61936]: DEBUG nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1004.655528] env[61936]: DEBUG nova.network.neutron [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1004.659501] env[61936]: DEBUG oslo_concurrency.lockutils [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.659501] env[61936]: DEBUG nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Received event network-vif-deleted-e21d453a-d68d-4497-9042-ebc1a0dbec83 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1004.659501] env[61936]: INFO nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Neutron deleted interface e21d453a-d68d-4497-9042-ebc1a0dbec83; detaching it from the instance and deleting it from the info cache [ 1004.659501] env[61936]: DEBUG nova.network.neutron [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.695187] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253329, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067661} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.695571] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.696558] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ca6e31-45d7-44b3-b07d-d2ac30ac15cc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.726872] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 755c5fea-e3d2-4cb3-b717-5f862401d7e1/755c5fea-e3d2-4cb3-b717-5f862401d7e1.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.729856] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05a1de54-d4e2-4077-a7fe-fa7a10f1b7c6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.751476] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1004.751476] env[61936]: value = "task-1253330" [ 1004.751476] env[61936]: _type = "Task" [ 1004.751476] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.762576] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253330, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.769879] env[61936]: DEBUG nova.policy [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9260b79bb7f4e44bb1f50b16273df09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ce9d789b30c49758bd073d17b3f1281', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.787018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.787018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.787018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.787018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.787018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.789317] env[61936]: INFO nova.compute.manager [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Terminating instance [ 1004.903530] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d91777-c52c-4b77-b09a-07fa5a1c63f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.911512] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc5e424-13ef-4c5f-9162-f78229f7647d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.950141] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83abbacd-afc1-45a0-bd1d-c733d52d70cd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.961431] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Preparing fetch location {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1004.961701] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Fetch image to [datastore2] OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d/OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d.vmdk {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1004.961889] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Downloading stream optimized image 2f04c09d-c433-442e-b891-3fa38ecc215e to [datastore2] OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d/OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d.vmdk on the data store datastore2 as vApp {{(pid=61936) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1004.962086] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Downloading image file data 2f04c09d-c433-442e-b891-3fa38ecc215e to the ESX as VM named 'OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d' {{(pid=61936) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1004.964520] env[61936]: DEBUG nova.network.neutron [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Successfully updated port: 855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.967765] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967fc0de-4eca-47d6-be54-9da5b96f90df {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.985708] env[61936]: DEBUG nova.compute.provider_tree [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.041994] env[61936]: DEBUG nova.compute.manager [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Received event network-vif-plugged-855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1005.042602] env[61936]: DEBUG oslo_concurrency.lockutils [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] Acquiring lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.042823] env[61936]: DEBUG oslo_concurrency.lockutils [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.042991] env[61936]: DEBUG oslo_concurrency.lockutils [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.043181] env[61936]: DEBUG nova.compute.manager [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] No waiting events found dispatching network-vif-plugged-855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1005.043369] env[61936]: WARNING nova.compute.manager [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Received unexpected event network-vif-plugged-855dd2af-6f16-4185-88e2-0022244dc30d for instance with vm_state building and task_state spawning. [ 1005.043688] env[61936]: DEBUG nova.compute.manager [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Received event network-changed-855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1005.043688] env[61936]: DEBUG nova.compute.manager [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Refreshing instance network info cache due to event network-changed-855dd2af-6f16-4185-88e2-0022244dc30d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1005.043840] env[61936]: DEBUG oslo_concurrency.lockutils [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] Acquiring lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.043976] env[61936]: DEBUG oslo_concurrency.lockutils [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] Acquired lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.044215] env[61936]: DEBUG nova.network.neutron [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Refreshing network info cache for port 855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1005.056832] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1005.056832] env[61936]: value = "resgroup-9" [ 1005.056832] env[61936]: _type = "ResourcePool" [ 1005.056832] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1005.057506] env[61936]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0208d4ff-9a67-4f89-b320-f7ad0cc6a477 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.078359] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lease: (returnval){ [ 1005.078359] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cd0325-9e63-1bdc-9072-f5241a150ea4" [ 1005.078359] env[61936]: _type = "HttpNfcLease" [ 1005.078359] env[61936]: } obtained for vApp import into resource pool (val){ [ 1005.078359] env[61936]: value = "resgroup-9" [ 1005.078359] env[61936]: _type = "ResourcePool" [ 1005.078359] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1005.078761] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the lease: (returnval){ [ 1005.078761] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cd0325-9e63-1bdc-9072-f5241a150ea4" [ 1005.078761] env[61936]: _type = "HttpNfcLease" [ 1005.078761] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1005.085789] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1005.085789] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cd0325-9e63-1bdc-9072-f5241a150ea4" [ 1005.085789] env[61936]: _type = "HttpNfcLease" [ 1005.085789] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1005.155508] env[61936]: DEBUG nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1005.161509] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-059830bd-1426-4423-b910-179c347dfa1a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.171282] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3802e316-bdd4-41b0-b726-33583963e0e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.202144] env[61936]: DEBUG nova.compute.manager [req-a7fb8006-c77b-4cf6-b5f7-1d7c0c7c0704 req-7bea0c5c-3bc2-40bb-9a56-49ee8f13e234 service nova] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Detach interface failed, port_id=e21d453a-d68d-4497-9042-ebc1a0dbec83, reason: Instance 90c2fd2c-937c-4b92-8bb1-6387652a03bb could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1005.261701] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253330, 'name': ReconfigVM_Task, 'duration_secs': 0.344012} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.262038] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 755c5fea-e3d2-4cb3-b717-5f862401d7e1/755c5fea-e3d2-4cb3-b717-5f862401d7e1.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.262810] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6086d2e0-123f-4d3d-a2f9-a8dbe5195b93 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.265771] env[61936]: DEBUG nova.network.neutron [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Successfully created port: b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.270262] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1005.270262] env[61936]: value = "task-1253332" [ 1005.270262] env[61936]: _type = "Task" [ 1005.270262] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.283453] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253332, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.293699] env[61936]: DEBUG nova.compute.manager [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1005.294220] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1005.295263] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54aeb7f2-36c0-4063-a8b4-f22e8c70317f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.302795] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.303099] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec55195b-52c8-4ef7-a11a-d6656abcd979 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.307425] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.307699] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eeb514fb-1763-4007-a3d5-b49ac717e0c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.311506] env[61936]: DEBUG oslo_vmware.api [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1005.311506] env[61936]: value = "task-1253333" [ 1005.311506] env[61936]: _type = "Task" [ 1005.311506] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.315904] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1005.315904] env[61936]: value = "task-1253334" [ 1005.315904] env[61936]: _type = "Task" [ 1005.315904] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.322825] env[61936]: DEBUG oslo_vmware.api [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253333, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.332772] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.476048] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.490673] env[61936]: DEBUG nova.scheduler.client.report [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1005.589054] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1005.589054] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cd0325-9e63-1bdc-9072-f5241a150ea4" [ 1005.589054] env[61936]: _type = "HttpNfcLease" [ 1005.589054] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1005.606924] env[61936]: DEBUG nova.network.neutron [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1005.780796] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253332, 'name': Rename_Task, 'duration_secs': 0.141769} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.781199] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.781355] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-753fc083-5460-4182-8d56-c7240e602f3e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.788213] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1005.788213] env[61936]: value = "task-1253335" [ 1005.788213] env[61936]: _type = "Task" [ 1005.788213] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.795868] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.823500] env[61936]: DEBUG oslo_vmware.api [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253333, 'name': PowerOffVM_Task, 'duration_secs': 0.186384} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.825135] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.825355] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1005.825610] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5f83cc5-7791-4bf8-808c-4ca3f4fa8537 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.830308] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253334, 'name': PowerOffVM_Task, 'duration_secs': 0.18329} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.830878] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.831664] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804468b0-4434-4390-8235-c4932470271d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.849371] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c6bcec-408d-4d99-a579-8f8012cb5885 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.892335] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1005.892576] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1005.892763] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Deleting the datastore file [datastore1] 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.893746] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ef73b29-978a-4540-aa86-08e83e4fa8b6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.900056] env[61936]: DEBUG oslo_vmware.api [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for the task: (returnval){ [ 1005.900056] env[61936]: value = "task-1253337" [ 1005.900056] env[61936]: _type = "Task" [ 1005.900056] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.907501] env[61936]: DEBUG oslo_vmware.api [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253337, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.995574] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.998540] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.159s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.998819] env[61936]: DEBUG nova.objects.instance [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'resources' on Instance uuid dc6523b7-c15c-469f-83a5-5c5dc5f2d230 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.021084] env[61936]: INFO nova.scheduler.client.report [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted allocations for instance 62b33dd9-e523-4309-94e6-59671fcfbaea [ 1006.029727] env[61936]: DEBUG nova.network.neutron [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.087920] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.087920] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cd0325-9e63-1bdc-9072-f5241a150ea4" [ 1006.087920] env[61936]: _type = "HttpNfcLease" [ 1006.087920] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1006.088236] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1006.088236] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cd0325-9e63-1bdc-9072-f5241a150ea4" [ 1006.088236] env[61936]: _type = "HttpNfcLease" [ 1006.088236] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1006.088969] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482c3440-1d30-4ff1-bb92-3aa158a473b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.096273] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281dfb3-895b-e247-93ef-c41a589b2c85/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1006.096455] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating HTTP connection to write to file with size = 31666688 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281dfb3-895b-e247-93ef-c41a589b2c85/disk-0.vmdk. {{(pid=61936) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1006.161342] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3bc77f74-1180-4150-962d-8f41f7d36ee5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.165972] env[61936]: DEBUG nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1006.195369] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1006.195659] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.195821] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1006.196037] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.196227] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1006.196428] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1006.196699] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1006.196938] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1006.197123] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1006.197315] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1006.197520] env[61936]: DEBUG nova.virt.hardware [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1006.198414] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99be61e8-f7c7-4bc4-9302-5a78bfa6a0f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.208360] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad53ef2-5f89-4a33-ad04-1945510d192f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.297445] env[61936]: DEBUG oslo_vmware.api [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253335, 'name': PowerOnVM_Task, 'duration_secs': 0.472107} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.297706] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.297909] env[61936]: INFO nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Took 8.08 seconds to spawn the instance on the hypervisor. [ 1006.298102] env[61936]: DEBUG nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1006.298846] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2e9774-bed3-4f51-95b9-50908e7cdd6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.359695] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Creating Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1006.360155] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e28f9f39-6921-4803-a795-1856ed1fc493 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.369144] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1006.369144] env[61936]: value = "task-1253338" [ 1006.369144] env[61936]: _type = "Task" [ 1006.369144] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.380809] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253338, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.412105] env[61936]: DEBUG oslo_vmware.api [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Task: {'id': task-1253337, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180343} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.414185] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.414341] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1006.414441] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1006.414606] env[61936]: INFO nova.compute.manager [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1006.414863] env[61936]: DEBUG oslo.service.loopingcall [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.415184] env[61936]: DEBUG nova.compute.manager [-] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1006.415279] env[61936]: DEBUG nova.network.neutron [-] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1006.504846] env[61936]: DEBUG nova.objects.instance [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'numa_topology' on Instance uuid dc6523b7-c15c-469f-83a5-5c5dc5f2d230 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.531210] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a05ce5cc-e80b-430f-84e6-6f57460a385e tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "62b33dd9-e523-4309-94e6-59671fcfbaea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.806s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.537324] env[61936]: DEBUG oslo_concurrency.lockutils [req-bbd4e6a3-02d4-486f-86e5-f14e6fae135b req-3b983729-5752-40eb-b8b5-47b8317ac51f service nova] Releasing lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.537730] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.537856] env[61936]: DEBUG nova.network.neutron [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.817141] env[61936]: INFO nova.compute.manager [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Took 24.82 seconds to build instance. [ 1006.882279] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253338, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.009856] env[61936]: DEBUG nova.objects.base [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1007.133415] env[61936]: DEBUG nova.network.neutron [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1007.195064] env[61936]: DEBUG nova.network.neutron [-] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.241676] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955eab80-96f7-4449-9184-e3104acd0d7e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.256621] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739c6dec-4a4f-4307-9f1b-e561193a22c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.301127] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f74f4db-0024-4289-a15a-7c93d7e24704 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.311139] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1ae149-15f4-42b8-94fc-d56458490e21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.319249] env[61936]: DEBUG nova.compute.manager [req-461a3683-4ba0-4ea0-aae1-477ad1bc3548 req-fed06eca-8d7b-4288-87d9-6570f12a92bd service nova] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Received event network-vif-deleted-4ecf3c38-78ee-498b-bcf6-8a98cffbcc1b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1007.320522] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be764ad5-af8e-4f6f-8684-abb4fd239f8d tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.339s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.335207] env[61936]: DEBUG nova.compute.provider_tree [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.349018] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Completed reading data from the image iterator. {{(pid=61936) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1007.349421] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281dfb3-895b-e247-93ef-c41a589b2c85/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1007.351033] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b324fb0-fa8f-4d55-aac7-b22d39031b42 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.360247] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281dfb3-895b-e247-93ef-c41a589b2c85/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1007.360493] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281dfb3-895b-e247-93ef-c41a589b2c85/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1007.360740] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1ea22762-1187-4bc4-a3e2-a976d3dfa1fe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.380928] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253338, 'name': CreateSnapshot_Task, 'duration_secs': 0.76684} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.380928] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Created Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1007.381823] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cfccd9-a88e-4e49-8fca-ec97c1ee3157 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.538280] env[61936]: DEBUG nova.network.neutron [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updating instance_info_cache with network_info: [{"id": "855dd2af-6f16-4185-88e2-0022244dc30d", "address": "fa:16:3e:b1:24:f2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855dd2af-6f", "ovs_interfaceid": "855dd2af-6f16-4185-88e2-0022244dc30d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.574296] env[61936]: DEBUG oslo_vmware.rw_handles [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281dfb3-895b-e247-93ef-c41a589b2c85/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1007.575620] env[61936]: INFO nova.virt.vmwareapi.images [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Downloaded image file data 2f04c09d-c433-442e-b891-3fa38ecc215e [ 1007.576206] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9cfae99-05cd-4cd5-acfe-e074b79e1f52 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.594576] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85ad5440-d369-4e8a-946d-518ef5d3e593 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.619095] env[61936]: INFO nova.virt.vmwareapi.images [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] The imported VM was unregistered [ 1007.621998] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Caching image {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1007.622143] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Creating directory with path [datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1007.622400] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e32a2827-60a4-4f42-8719-4838aaa938a0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.648670] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Created directory with path [datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1007.648994] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d/OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d.vmdk to [datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e/2f04c09d-c433-442e-b891-3fa38ecc215e.vmdk. {{(pid=61936) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1007.649298] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-fa702788-bdab-4d29-968d-c72c0ed74a73 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.656047] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1007.656047] env[61936]: value = "task-1253340" [ 1007.656047] env[61936]: _type = "Task" [ 1007.656047] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.666715] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253340, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.666715] env[61936]: DEBUG nova.network.neutron [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Successfully updated port: b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.697087] env[61936]: INFO nova.compute.manager [-] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Took 1.28 seconds to deallocate network for instance. [ 1007.858803] env[61936]: DEBUG nova.scheduler.client.report [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1007.903144] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Creating linked-clone VM from snapshot {{(pid=61936) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1007.904694] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1d00c6d6-1415-4037-a582-5418daa1f774 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.913510] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1007.913510] env[61936]: value = "task-1253341" [ 1007.913510] env[61936]: _type = "Task" [ 1007.913510] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.922482] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253341, 'name': CloneVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.041024] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.041523] env[61936]: DEBUG nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Instance network_info: |[{"id": "855dd2af-6f16-4185-88e2-0022244dc30d", "address": "fa:16:3e:b1:24:f2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855dd2af-6f", "ovs_interfaceid": "855dd2af-6f16-4185-88e2-0022244dc30d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1008.042096] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:24:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '855dd2af-6f16-4185-88e2-0022244dc30d', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.050148] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating folder: Project (aa564b684410493fa0028fd345048e02). Parent ref: group-v269874. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1008.050968] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35c05e3d-6559-4ee5-a0e5-fe6896636f76 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.063731] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created folder: Project (aa564b684410493fa0028fd345048e02) in parent group-v269874. [ 1008.063990] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating folder: Instances. Parent ref: group-v270016. {{(pid=61936) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1008.064283] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f3c1130-2d26-4b85-b066-cda9ccb68d80 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.079029] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created folder: Instances in parent group-v270016. [ 1008.079029] env[61936]: DEBUG oslo.service.loopingcall [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.079029] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1008.079029] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22acccee-866d-40e0-a1d7-e33a8824339d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.098787] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.098787] env[61936]: value = "task-1253344" [ 1008.098787] env[61936]: _type = "Task" [ 1008.098787] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.110210] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253344, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.173336] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.173733] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.174071] env[61936]: DEBUG nova.network.neutron [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.175510] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253340, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.203397] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.330833] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "b780470b-57d0-43e5-be52-539c78cada6b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.331127] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "b780470b-57d0-43e5-be52-539c78cada6b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.331351] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "b780470b-57d0-43e5-be52-539c78cada6b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.331536] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "b780470b-57d0-43e5-be52-539c78cada6b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.331714] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "b780470b-57d0-43e5-be52-539c78cada6b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.335018] env[61936]: INFO nova.compute.manager [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Terminating instance [ 1008.362381] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.364s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.365565] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.925s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.365817] env[61936]: DEBUG nova.objects.instance [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lazy-loading 'resources' on Instance uuid 90c2fd2c-937c-4b92-8bb1-6387652a03bb {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.427044] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253341, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.618110] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253344, 'name': CreateVM_Task, 'duration_secs': 0.428915} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.618508] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1008.619802] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.620138] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.620673] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.621278] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f2c981b-e2be-434e-96f5-33ca221581c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.629585] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1008.629585] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e66c67-45d5-c4e6-85cc-c8630eb0aaf4" [ 1008.629585] env[61936]: _type = "Task" [ 1008.629585] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.643409] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e66c67-45d5-c4e6-85cc-c8630eb0aaf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.669734] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253340, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.709425] env[61936]: DEBUG nova.network.neutron [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.854990] env[61936]: DEBUG nova.compute.manager [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1008.854990] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.856213] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64b048d-b10c-4a9d-86e5-105b794ac295 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.873774] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.876704] env[61936]: DEBUG oslo_concurrency.lockutils [None req-23256108-6f5e-4d35-ad5c-f18e1db81d04 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 31.389s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.877749] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12c70fcc-8bae-4dde-b10e-7301e94c38fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.881176] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 9.893s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.881401] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.881701] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.883481] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.885063] env[61936]: INFO nova.compute.manager [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Terminating instance [ 1008.894443] env[61936]: DEBUG oslo_vmware.api [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1008.894443] env[61936]: value = "task-1253345" [ 1008.894443] env[61936]: _type = "Task" [ 1008.894443] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.920345] env[61936]: DEBUG oslo_vmware.api [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253345, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.939491] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253341, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.039273] env[61936]: DEBUG nova.network.neutron [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updating instance_info_cache with network_info: [{"id": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "address": "fa:16:3e:44:81:ac", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb556882a-8f", "ovs_interfaceid": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.148078] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e66c67-45d5-c4e6-85cc-c8630eb0aaf4, 'name': SearchDatastore_Task, 'duration_secs': 0.085045} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.148078] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.148305] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.148483] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.148682] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.148888] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.152058] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa751c0a-50e2-4c34-8df8-8fa4e19edb5c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.173879] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "1c391c45-a041-422d-9a44-a29306f99a6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.174140] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "1c391c45-a041-422d-9a44-a29306f99a6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.175324] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253340, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.177397] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.177571] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.179155] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0ab0965-487b-469f-aada-e851c418016f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.188043] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1009.188043] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f91e6f-2c44-2a3e-60e5-f2c39fec3343" [ 1009.188043] env[61936]: _type = "Task" [ 1009.188043] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.194735] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b6a252-e3c7-4230-a744-42702d1518d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.201234] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f91e6f-2c44-2a3e-60e5-f2c39fec3343, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.205567] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ab9919-28d8-46e3-8fac-6d2cdce2359f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.239427] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ff8840-ae20-47e5-a317-b8c018405071 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.248984] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98aec75d-0058-4f10-8d4d-ad2b6052b5b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.263205] env[61936]: DEBUG nova.compute.provider_tree [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.347044] env[61936]: DEBUG nova.compute.manager [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Received event network-vif-plugged-b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1009.347337] env[61936]: DEBUG oslo_concurrency.lockutils [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] Acquiring lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.347563] env[61936]: DEBUG oslo_concurrency.lockutils [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.347736] env[61936]: DEBUG oslo_concurrency.lockutils [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.347905] env[61936]: DEBUG nova.compute.manager [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] No waiting events found dispatching network-vif-plugged-b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1009.348425] env[61936]: WARNING nova.compute.manager [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Received unexpected event network-vif-plugged-b556882a-8fa4-4ea2-871b-6784ef87c88c for instance with vm_state building and task_state spawning. [ 1009.348528] env[61936]: DEBUG nova.compute.manager [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Received event network-changed-b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1009.348696] env[61936]: DEBUG nova.compute.manager [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Refreshing instance network info cache due to event network-changed-b556882a-8fa4-4ea2-871b-6784ef87c88c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1009.348881] env[61936]: DEBUG oslo_concurrency.lockutils [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] Acquiring lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.389176] env[61936]: DEBUG nova.compute.manager [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1009.389473] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.389981] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e00df76-e3b3-43dd-bf80-8609d2c58e53 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.399168] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0b6d46-1039-4db1-9621-56fdb978db70 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.420057] env[61936]: DEBUG nova.compute.manager [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Stashing vm_state: active {{(pid=61936) _prep_resize /opt/stack/nova/nova/compute/manager.py:6015}} [ 1009.440139] env[61936]: DEBUG oslo_vmware.api [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253345, 'name': PowerOffVM_Task, 'duration_secs': 0.409261} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.440335] env[61936]: WARNING nova.virt.vmwareapi.vmops [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc6523b7-c15c-469f-83a5-5c5dc5f2d230 could not be found. [ 1009.440507] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.440691] env[61936]: INFO nova.compute.manager [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1009.440976] env[61936]: DEBUG oslo.service.loopingcall [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.441568] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.441763] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1009.443174] env[61936]: DEBUG nova.compute.manager [-] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1009.443174] env[61936]: DEBUG nova.network.neutron [-] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.444339] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eeb43971-bd87-4e61-9172-16fa64b87b18 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.449513] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253341, 'name': CloneVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.512483] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1009.512760] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1009.512912] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleting the datastore file [datastore1] b780470b-57d0-43e5-be52-539c78cada6b {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.513224] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34bfeccb-69c8-4387-b4c8-b0218fd15827 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.523921] env[61936]: DEBUG oslo_vmware.api [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1009.523921] env[61936]: value = "task-1253347" [ 1009.523921] env[61936]: _type = "Task" [ 1009.523921] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.537295] env[61936]: DEBUG oslo_vmware.api [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253347, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.545073] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.545928] env[61936]: DEBUG nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Instance network_info: |[{"id": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "address": "fa:16:3e:44:81:ac", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb556882a-8f", "ovs_interfaceid": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1009.546513] env[61936]: DEBUG oslo_concurrency.lockutils [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] Acquired lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.546851] env[61936]: DEBUG nova.network.neutron [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Refreshing network info cache for port b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.549552] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:81:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b556882a-8fa4-4ea2-871b-6784ef87c88c', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.558890] env[61936]: DEBUG oslo.service.loopingcall [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.559882] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.560175] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca4e59cf-6489-4fd5-b46a-9ed80a68a294 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.582728] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.582728] env[61936]: value = "task-1253348" [ 1009.582728] env[61936]: _type = "Task" [ 1009.582728] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.595206] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253348, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.674952] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253340, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.679171] env[61936]: DEBUG nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1009.702485] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f91e6f-2c44-2a3e-60e5-f2c39fec3343, 'name': SearchDatastore_Task, 'duration_secs': 0.083227} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.702800] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80d2199e-9ba5-43ac-8d5b-71ed34e92545 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.713467] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1009.713467] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cedb1a-a693-de26-f503-013a90b8ab87" [ 1009.713467] env[61936]: _type = "Task" [ 1009.713467] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.727148] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cedb1a-a693-de26-f503-013a90b8ab87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.766286] env[61936]: DEBUG nova.scheduler.client.report [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1009.937596] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253341, 'name': CloneVM_Task, 'duration_secs': 1.559002} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.938037] env[61936]: INFO nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Created linked-clone VM from snapshot [ 1009.938795] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecee5124-69e7-4a14-b545-a46c9a2f9470 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.952551] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Uploading image baa3a86e-79f6-4c48-b14b-11d628f6dec2 {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1009.960346] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.980781] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1009.980781] env[61936]: value = "vm-270015" [ 1009.980781] env[61936]: _type = "VirtualMachine" [ 1009.980781] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1009.981304] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-34079732-2625-49e8-881f-ee26590db1f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.990824] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease: (returnval){ [ 1009.990824] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524367b2-48e5-5238-3c57-1184c1bff96e" [ 1009.990824] env[61936]: _type = "HttpNfcLease" [ 1009.990824] env[61936]: } obtained for exporting VM: (result){ [ 1009.990824] env[61936]: value = "vm-270015" [ 1009.990824] env[61936]: _type = "VirtualMachine" [ 1009.990824] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1009.991312] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the lease: (returnval){ [ 1009.991312] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524367b2-48e5-5238-3c57-1184c1bff96e" [ 1009.991312] env[61936]: _type = "HttpNfcLease" [ 1009.991312] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1010.002544] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1010.002544] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524367b2-48e5-5238-3c57-1184c1bff96e" [ 1010.002544] env[61936]: _type = "HttpNfcLease" [ 1010.002544] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1010.040331] env[61936]: DEBUG oslo_vmware.api [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253347, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32783} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.040820] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1010.041145] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1010.041473] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1010.041789] env[61936]: INFO nova.compute.manager [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1010.042392] env[61936]: DEBUG oslo.service.loopingcall [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.042546] env[61936]: DEBUG nova.compute.manager [-] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1010.042726] env[61936]: DEBUG nova.network.neutron [-] [instance: b780470b-57d0-43e5-be52-539c78cada6b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1010.094560] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253348, 'name': CreateVM_Task, 'duration_secs': 0.390504} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.094790] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1010.095476] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.095647] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.095982] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.096269] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89e8b6f3-91f1-47b1-b14d-e048b2fb46d7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.102948] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1010.102948] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e4e3ad-c297-85f4-7df1-dba7e06a24b0" [ 1010.102948] env[61936]: _type = "Task" [ 1010.102948] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.113888] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e4e3ad-c297-85f4-7df1-dba7e06a24b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.173780] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253340, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.200218] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.226821] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52cedb1a-a693-de26-f503-013a90b8ab87, 'name': SearchDatastore_Task, 'duration_secs': 0.100919} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.227498] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.227804] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] e6930eaf-bb63-4553-8cd2-f2eb62fca510/e6930eaf-bb63-4553-8cd2-f2eb62fca510.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.228100] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3a2536d-0f5f-4c3e-8e0a-10706e086ab9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.238605] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1010.238605] env[61936]: value = "task-1253350" [ 1010.238605] env[61936]: _type = "Task" [ 1010.238605] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.250009] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.272330] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.274656] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.072s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.274894] env[61936]: DEBUG nova.objects.instance [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lazy-loading 'resources' on Instance uuid 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.294731] env[61936]: INFO nova.scheduler.client.report [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted allocations for instance 90c2fd2c-937c-4b92-8bb1-6387652a03bb [ 1010.317626] env[61936]: DEBUG nova.network.neutron [-] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.499720] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1010.499720] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524367b2-48e5-5238-3c57-1184c1bff96e" [ 1010.499720] env[61936]: _type = "HttpNfcLease" [ 1010.499720] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1010.500137] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1010.500137] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]524367b2-48e5-5238-3c57-1184c1bff96e" [ 1010.500137] env[61936]: _type = "HttpNfcLease" [ 1010.500137] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1010.500987] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb819c23-0739-4b67-84a5-25e85db5c88a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.508317] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527f2026-0c74-5e37-fd69-56edf94edd72/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1010.508499] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527f2026-0c74-5e37-fd69-56edf94edd72/disk-0.vmdk for reading. {{(pid=61936) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1010.601524] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f33ce838-3353-41f4-a9df-01e3eab6cfb8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.614893] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e4e3ad-c297-85f4-7df1-dba7e06a24b0, 'name': SearchDatastore_Task, 'duration_secs': 0.025202} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.617295] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.617585] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.617877] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.618076] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.618298] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.619100] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08c05d4d-fdbf-416f-80fc-a9d900f0ca3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.628856] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.629084] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1010.629870] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc7fc0fd-c1e2-4718-8f69-7451195663ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.639065] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1010.639065] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c1df57-8bb4-d0eb-d5ae-53447807423b" [ 1010.639065] env[61936]: _type = "Task" [ 1010.639065] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.647899] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c1df57-8bb4-d0eb-d5ae-53447807423b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.673202] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253340, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.963703} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.673202] env[61936]: INFO nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d/OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d.vmdk to [datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e/2f04c09d-c433-442e-b891-3fa38ecc215e.vmdk. [ 1010.673202] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Cleaning up location [datastore2] OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1010.673202] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_2a01f2f1-eded-472d-b98a-c467a4f0b68d {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.673202] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f214bd0-9407-4ea2-981d-b32f55156ddf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.678702] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1010.678702] env[61936]: value = "task-1253351" [ 1010.678702] env[61936]: _type = "Task" [ 1010.678702] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.688759] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253351, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.702396] env[61936]: DEBUG nova.network.neutron [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updated VIF entry in instance network info cache for port b556882a-8fa4-4ea2-871b-6784ef87c88c. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.702754] env[61936]: DEBUG nova.network.neutron [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updating instance_info_cache with network_info: [{"id": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "address": "fa:16:3e:44:81:ac", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb556882a-8f", "ovs_interfaceid": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.756697] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253350, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.809049] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bca4188-64b0-472b-a148-4207675b3694 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "90c2fd2c-937c-4b92-8bb1-6387652a03bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.212s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.819597] env[61936]: INFO nova.compute.manager [-] [instance: dc6523b7-c15c-469f-83a5-5c5dc5f2d230] Took 1.38 seconds to deallocate network for instance. [ 1011.035464] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80d6d5b-a900-4cb0-ad6e-82a6953b6b30 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.046059] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebb496b-9e30-4f8d-a2f7-0c6eba4206b3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.078569] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ea75bd-fabc-4b2c-8fd5-2d7462931756 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.087983] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de7ce1e-30f6-4d07-b548-274d71a34b0f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.102539] env[61936]: DEBUG nova.compute.provider_tree [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.148500] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c1df57-8bb4-d0eb-d5ae-53447807423b, 'name': SearchDatastore_Task, 'duration_secs': 0.03769} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.149531] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4460df31-acbb-4c02-b910-8b0206e432ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.155378] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1011.155378] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521e6faf-6264-5fcd-bdb9-5d6d86667fd0" [ 1011.155378] env[61936]: _type = "Task" [ 1011.155378] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.164466] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521e6faf-6264-5fcd-bdb9-5d6d86667fd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.185738] env[61936]: DEBUG nova.network.neutron [-] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.192219] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253351, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175892} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.192760] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.192949] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e/2f04c09d-c433-442e-b891-3fa38ecc215e.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.193288] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e/2f04c09d-c433-442e-b891-3fa38ecc215e.vmdk to [datastore2] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.193623] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d723cbb-275a-46e7-bb30-92d15267d4de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.201848] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1011.201848] env[61936]: value = "task-1253352" [ 1011.201848] env[61936]: _type = "Task" [ 1011.201848] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.206461] env[61936]: DEBUG oslo_concurrency.lockutils [req-36eb52dd-7cbd-47b3-adcf-2d46e9b5a8cf req-446d9447-1f91-42f5-a151-7f763984aae1 service nova] Releasing lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.213008] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253352, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.250158] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253350, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.376965] env[61936]: DEBUG nova.compute.manager [req-92bb031d-0ed5-4b7a-b6d7-105e048b44de req-28af379e-6888-4945-bac1-7ef7bb7b770a service nova] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Received event network-vif-deleted-8452d95f-2741-4d7d-a681-67321259085c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1011.399900] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.400569] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.401529] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.401529] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.401903] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.404787] env[61936]: INFO nova.compute.manager [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Terminating instance [ 1011.606187] env[61936]: DEBUG nova.scheduler.client.report [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1011.669071] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521e6faf-6264-5fcd-bdb9-5d6d86667fd0, 'name': SearchDatastore_Task, 'duration_secs': 0.028712} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.669071] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.669316] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b/18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.669628] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6c04cc3-bef2-4c57-9ae6-137a367591eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.676630] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1011.676630] env[61936]: value = "task-1253353" [ 1011.676630] env[61936]: _type = "Task" [ 1011.676630] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.684251] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.693211] env[61936]: INFO nova.compute.manager [-] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Took 1.65 seconds to deallocate network for instance. [ 1011.714892] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253352, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.752256] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253350, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.114122} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.752684] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] e6930eaf-bb63-4553-8cd2-f2eb62fca510/e6930eaf-bb63-4553-8cd2-f2eb62fca510.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.752989] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.753597] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2f062a7-d113-48cc-82f1-0a1722b16999 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.760764] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1011.760764] env[61936]: value = "task-1253354" [ 1011.760764] env[61936]: _type = "Task" [ 1011.760764] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.769409] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253354, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.856251] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae0e3019-9729-475d-b426-28c4ad2e0ee3 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "dc6523b7-c15c-469f-83a5-5c5dc5f2d230" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.975s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.910133] env[61936]: DEBUG nova.compute.manager [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1011.910488] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1011.912413] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2e2e30-fd24-4529-bd85-39eaeb86c592 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.918750] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.919269] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.924707] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.925537] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b5e2ea4-305d-401e-bad4-7e7665dab1fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.934650] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1011.934650] env[61936]: value = "task-1253355" [ 1011.934650] env[61936]: _type = "Task" [ 1011.934650] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.948880] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.112365] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.116014] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.155s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.145468] env[61936]: INFO nova.scheduler.client.report [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Deleted allocations for instance 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6 [ 1012.188722] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253353, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.200358] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.214533] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253352, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.272995] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253354, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.425605} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.275586] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.275586] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16a0b0f-05d0-4d84-b6ac-ae398e80c70d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.300350] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] e6930eaf-bb63-4553-8cd2-f2eb62fca510/e6930eaf-bb63-4553-8cd2-f2eb62fca510.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.300820] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0eb2624d-3aca-4590-95e3-c76a2db18a5d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.322318] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1012.322318] env[61936]: value = "task-1253356" [ 1012.322318] env[61936]: _type = "Task" [ 1012.322318] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.331071] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253356, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.426592] env[61936]: DEBUG nova.compute.utils [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1012.448831] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253355, 'name': PowerOffVM_Task, 'duration_secs': 0.223248} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.449255] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.449462] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1012.449742] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb0b0586-c0d4-4919-a011-780e8a79d41b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.555815] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.556173] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.556504] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleting the datastore file [datastore2] 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.556891] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fbcab4a-64db-4ffa-987d-1481627a7a56 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.564443] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1012.564443] env[61936]: value = "task-1253358" [ 1012.564443] env[61936]: _type = "Task" [ 1012.564443] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.573346] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.621679] env[61936]: INFO nova.compute.claims [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.655391] env[61936]: DEBUG oslo_concurrency.lockutils [None req-186c52b0-a856-43dc-80c1-1998a87ebc44 tempest-ServerAddressesNegativeTestJSON-1100005245 tempest-ServerAddressesNegativeTestJSON-1100005245-project-member] Lock "8b6696e5-ef1b-4825-903f-de4dfb9bf2a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.870s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.688842] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.853116} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.689255] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b/18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1012.689499] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.689802] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50fd821c-6e78-4e06-965c-8607ec418140 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.697484] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1012.697484] env[61936]: value = "task-1253359" [ 1012.697484] env[61936]: _type = "Task" [ 1012.697484] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.716548] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253359, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.720208] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253352, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.835851] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253356, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.930396] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.078096] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.128764] env[61936]: INFO nova.compute.resource_tracker [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating resource usage from migration 2d0e61d3-548f-4a27-a434-95d93432350b [ 1013.212865] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.213234] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.215126] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253359, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078041} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.224251] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.225245] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b386c1-6a14-4732-a8a9-ec526f4cc3a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.252986] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b/18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.259454] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a8a9976-0790-4314-8b91-063e776f21af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.273903] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253352, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.285566] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1013.285566] env[61936]: value = "task-1253360" [ 1013.285566] env[61936]: _type = "Task" [ 1013.285566] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.302952] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253360, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.335382] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253356, 'name': ReconfigVM_Task, 'duration_secs': 0.542381} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.338564] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Reconfigured VM instance instance-0000005f to attach disk [datastore2] e6930eaf-bb63-4553-8cd2-f2eb62fca510/e6930eaf-bb63-4553-8cd2-f2eb62fca510.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.339641] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77d92230-3a72-4d3e-8989-67bcb5225212 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.350026] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1013.350026] env[61936]: value = "task-1253361" [ 1013.350026] env[61936]: _type = "Task" [ 1013.350026] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.365209] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253361, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.431195] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6937b3-132e-484f-9120-612243a80dd0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.442773] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0801603f-00d5-4f47-ba62-3c906b6034c1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.481195] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd6f702-470a-4d9b-a964-9be94d89e006 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.493538] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64132e2-cabb-45dd-9b4d-9398222dea33 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.512964] env[61936]: DEBUG nova.compute.provider_tree [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.581591] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.717344] env[61936]: DEBUG nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1013.731219] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253352, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.801432] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253360, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.872953] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253361, 'name': Rename_Task, 'duration_secs': 0.264772} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.872953] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.873573] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba346e95-52c6-4e39-9d6d-19d48217777a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.883373] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1013.883373] env[61936]: value = "task-1253362" [ 1013.883373] env[61936]: _type = "Task" [ 1013.883373] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.899522] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.013295] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.013578] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.013898] env[61936]: INFO nova.compute.manager [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Attaching volume 1d559b9c-feac-4939-85bd-15d6203183b4 to /dev/sdb [ 1014.020489] env[61936]: DEBUG nova.scheduler.client.report [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1014.059818] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e044d59-6b9e-419b-ae1a-99e6fbadf4ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.071623] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b826f90c-e579-4a0c-a40f-277f7acfc3c4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.080186] env[61936]: DEBUG oslo_vmware.api [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253358, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.494634} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.080454] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.080637] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.080813] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.080986] env[61936]: INFO nova.compute.manager [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Took 2.17 seconds to destroy the instance on the hypervisor. [ 1014.081251] env[61936]: DEBUG oslo.service.loopingcall [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.081519] env[61936]: DEBUG nova.compute.manager [-] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1014.081620] env[61936]: DEBUG nova.network.neutron [-] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1014.086653] env[61936]: DEBUG nova.virt.block_device [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating existing volume attachment record: e4253e64-3c80-490a-8f65-07dca3542d3c {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1014.227087] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253352, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.815917} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.229498] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2f04c09d-c433-442e-b891-3fa38ecc215e/2f04c09d-c433-442e-b891-3fa38ecc215e.vmdk to [datastore2] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1014.231524] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b47d805-cdba-4ff8-91c8-41f1862e12bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.253274] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.254633] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.254796] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ea725f2-ff36-453c-be43-92f8a54b5351 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.276308] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1014.276308] env[61936]: value = "task-1253364" [ 1014.276308] env[61936]: _type = "Task" [ 1014.276308] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.285274] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253364, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.297692] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253360, 'name': ReconfigVM_Task, 'duration_secs': 0.54662} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.297986] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b/18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.298655] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5bee81a-130f-4469-9898-3d7e98171e2d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.304860] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1014.304860] env[61936]: value = "task-1253365" [ 1014.304860] env[61936]: _type = "Task" [ 1014.304860] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.313861] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253365, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.394700] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253362, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.531332] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.416s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.531663] env[61936]: INFO nova.compute.manager [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Migrating [ 1014.539693] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.340s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.541422] env[61936]: INFO nova.compute.claims [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.766659] env[61936]: DEBUG nova.compute.manager [req-1b381185-27aa-4705-aedb-10d444daed56 req-43872029-a7f2-419b-95ef-ce65fadfe22f service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Received event network-vif-deleted-a3e152e9-3689-4f3d-ab48-c5f605ca1023 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1014.766931] env[61936]: INFO nova.compute.manager [req-1b381185-27aa-4705-aedb-10d444daed56 req-43872029-a7f2-419b-95ef-ce65fadfe22f service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Neutron deleted interface a3e152e9-3689-4f3d-ab48-c5f605ca1023; detaching it from the instance and deleting it from the info cache [ 1014.767177] env[61936]: DEBUG nova.network.neutron [req-1b381185-27aa-4705-aedb-10d444daed56 req-43872029-a7f2-419b-95ef-ce65fadfe22f service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.790422] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.814136] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253365, 'name': Rename_Task, 'duration_secs': 0.170847} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.814546] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.814858] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51e1b6c3-2cf0-41ab-8a0a-18506bd04a96 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.821514] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1014.821514] env[61936]: value = "task-1253368" [ 1014.821514] env[61936]: _type = "Task" [ 1014.821514] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.830917] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253368, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.894505] env[61936]: DEBUG oslo_vmware.api [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253362, 'name': PowerOnVM_Task, 'duration_secs': 0.553076} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.894725] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.895531] env[61936]: INFO nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Took 11.06 seconds to spawn the instance on the hypervisor. [ 1014.895531] env[61936]: DEBUG nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1014.896021] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84feaba8-c2c7-4b24-b4e0-1100b372174c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.058802] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.058802] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.058802] env[61936]: DEBUG nova.network.neutron [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.103486] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bf4e8f7d-2d21-4709-b678-9f8e6d465efe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-0137305a-dace-4eda-9d90-7233c001176f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.103486] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bf4e8f7d-2d21-4709-b678-9f8e6d465efe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.103772] env[61936]: DEBUG nova.objects.instance [None req-bf4e8f7d-2d21-4709-b678-9f8e6d465efe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'flavor' on Instance uuid 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.207983] env[61936]: DEBUG nova.network.neutron [-] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.269733] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5f89766-5399-4ae3-bb52-fb527d3e9ec5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.284192] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988cd450-1fa6-4809-a97f-df512ba5fd3e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.301737] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253364, 'name': ReconfigVM_Task, 'duration_secs': 0.659837} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.301982] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 5526cbd5-b1ad-453b-8401-eee7aa356606/5526cbd5-b1ad-453b-8401-eee7aa356606.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.302758] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4983de6a-8b6e-408a-8383-191cb000d16b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.314901] env[61936]: DEBUG nova.compute.manager [req-1b381185-27aa-4705-aedb-10d444daed56 req-43872029-a7f2-419b-95ef-ce65fadfe22f service nova] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Detach interface failed, port_id=a3e152e9-3689-4f3d-ab48-c5f605ca1023, reason: Instance 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1015.316607] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1015.316607] env[61936]: value = "task-1253369" [ 1015.316607] env[61936]: _type = "Task" [ 1015.316607] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.325603] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253369, 'name': Rename_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.336038] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253368, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.418180] env[61936]: INFO nova.compute.manager [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Took 20.67 seconds to build instance. [ 1015.607962] env[61936]: DEBUG nova.objects.instance [None req-bf4e8f7d-2d21-4709-b678-9f8e6d465efe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'pci_requests' on Instance uuid 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.713211] env[61936]: INFO nova.compute.manager [-] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Took 1.63 seconds to deallocate network for instance. [ 1015.802106] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfad236-35a4-47c6-abbe-bbdc3875075d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.810905] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00170d82-b87a-4b8a-902e-b7221454b5a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.855268] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7104092e-2bd6-499d-97c9-89a1443c8ed8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.862652] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253369, 'name': Rename_Task, 'duration_secs': 0.234153} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.866102] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.866102] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f52eeb23-12f4-4162-a166-91f1162cca14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.872721] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253368, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.874401] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e42441e-dcf1-4789-ad6e-d9657fe80213 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.879588] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1015.879588] env[61936]: value = "task-1253370" [ 1015.879588] env[61936]: _type = "Task" [ 1015.879588] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.890668] env[61936]: DEBUG nova.compute.provider_tree [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.898023] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253370, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.923820] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177548b0-d6ed-4684-9e50-00f016ceedcd tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.180s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.938273] env[61936]: DEBUG nova.network.neutron [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance_info_cache with network_info: [{"id": "7775677e-cf35-458b-8c94-da7dd8dc652a", "address": "fa:16:3e:f7:93:44", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7775677e-cf", "ovs_interfaceid": "7775677e-cf35-458b-8c94-da7dd8dc652a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.113978] env[61936]: DEBUG nova.objects.base [None req-bf4e8f7d-2d21-4709-b678-9f8e6d465efe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Object Instance<0137305a-dace-4eda-9d90-7233c001176f> lazy-loaded attributes: flavor,pci_requests {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1016.113978] env[61936]: DEBUG nova.network.neutron [None req-bf4e8f7d-2d21-4709-b678-9f8e6d465efe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1016.224407] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.300786] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bf4e8f7d-2d21-4709-b678-9f8e6d465efe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.197s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.358266] env[61936]: DEBUG oslo_vmware.api [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253368, 'name': PowerOnVM_Task, 'duration_secs': 1.397885} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.358565] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.358783] env[61936]: INFO nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Took 10.19 seconds to spawn the instance on the hypervisor. [ 1016.358954] env[61936]: DEBUG nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1016.359907] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd897528-e2be-4d70-9e8f-28c2e0b3b3de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.391062] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253370, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.393669] env[61936]: DEBUG nova.scheduler.client.report [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1016.442272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.809223] env[61936]: DEBUG nova.compute.manager [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Received event network-changed-855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1016.809428] env[61936]: DEBUG nova.compute.manager [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Refreshing instance network info cache due to event network-changed-855dd2af-6f16-4185-88e2-0022244dc30d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1016.809648] env[61936]: DEBUG oslo_concurrency.lockutils [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] Acquiring lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.809792] env[61936]: DEBUG oslo_concurrency.lockutils [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] Acquired lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.809950] env[61936]: DEBUG nova.network.neutron [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Refreshing network info cache for port 855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1016.878073] env[61936]: INFO nova.compute.manager [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Took 21.40 seconds to build instance. [ 1016.890768] env[61936]: DEBUG oslo_vmware.api [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253370, 'name': PowerOnVM_Task, 'duration_secs': 0.683514} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.891491] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.898852] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.899599] env[61936]: DEBUG nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1016.902663] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.702s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.902932] env[61936]: DEBUG nova.objects.instance [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lazy-loading 'resources' on Instance uuid b780470b-57d0-43e5-be52-539c78cada6b {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.998217] env[61936]: DEBUG nova.compute.manager [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1016.999479] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d95d7dc-2cc9-4366-8dac-3bd054e2e433 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.380804] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3a09cb3f-fcc1-4d7c-a806-bd84a86c6055 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.914s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.407033] env[61936]: DEBUG nova.compute.utils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1017.411738] env[61936]: DEBUG nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1017.412835] env[61936]: DEBUG nova.network.neutron [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1017.521293] env[61936]: DEBUG oslo_concurrency.lockutils [None req-00738d0e-ace6-4f8e-ab90-6dc5fe883c31 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.797s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.524882] env[61936]: DEBUG nova.policy [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b98c1bbe1044663830b18c4ce9e23d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1b7d01ffac74c6fa2306b4839041fdd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1017.528404] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1017.528404] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1017.676900] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9395a8c-db82-4be4-91ca-7e50069dd21a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.685770] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b0b01-7584-4a15-bf65-f661c6970890 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.719785] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725b1d0c-ac22-418b-aa31-27e37284ccf8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.727973] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e246027b-a0e0-471a-8a55-938c874f0a6f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.744025] env[61936]: DEBUG nova.compute.provider_tree [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.823807] env[61936]: DEBUG nova.network.neutron [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updated VIF entry in instance network info cache for port 855dd2af-6f16-4185-88e2-0022244dc30d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1017.824207] env[61936]: DEBUG nova.network.neutron [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updating instance_info_cache with network_info: [{"id": "855dd2af-6f16-4185-88e2-0022244dc30d", "address": "fa:16:3e:b1:24:f2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855dd2af-6f", "ovs_interfaceid": "855dd2af-6f16-4185-88e2-0022244dc30d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.913150] env[61936]: DEBUG nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1017.962025] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85636283-8dd3-487c-a7ab-63d5bc65c032 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.984927] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance '755c5fea-e3d2-4cb3-b717-5f862401d7e1' progress to 0 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1018.031777] env[61936]: DEBUG nova.network.neutron [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Successfully created port: a424508f-87c8-48b0-b5de-4a4916cddc3f {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.040810] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1018.040810] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 1018.246996] env[61936]: DEBUG nova.scheduler.client.report [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1018.272577] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-0137305a-dace-4eda-9d90-7233c001176f-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.272900] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.273417] env[61936]: DEBUG nova.objects.instance [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'flavor' on Instance uuid 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.327660] env[61936]: DEBUG oslo_concurrency.lockutils [req-38b01556-8031-4b50-bdcf-15bbe1851f43 req-78b6dc66-1e3e-48a2-a1a9-c94d284bc380 service nova] Releasing lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.433271] env[61936]: DEBUG nova.compute.manager [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Received event network-changed-b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1018.433574] env[61936]: DEBUG nova.compute.manager [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Refreshing instance network info cache due to event network-changed-b556882a-8fa4-4ea2-871b-6784ef87c88c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1018.433778] env[61936]: DEBUG oslo_concurrency.lockutils [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] Acquiring lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.433891] env[61936]: DEBUG oslo_concurrency.lockutils [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] Acquired lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.437233] env[61936]: DEBUG nova.network.neutron [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Refreshing network info cache for port b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.494870] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.495629] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ce2c26c-3636-4509-b3f5-17127f34e120 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.503493] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1018.503493] env[61936]: value = "task-1253372" [ 1018.503493] env[61936]: _type = "Task" [ 1018.503493] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.512627] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253372, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.580114] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.580249] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.580475] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Forcefully refreshing network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1018.636700] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1018.637047] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270021', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'name': 'volume-1d559b9c-feac-4939-85bd-15d6203183b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8', 'attached_at': '', 'detached_at': '', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'serial': '1d559b9c-feac-4939-85bd-15d6203183b4'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1018.638475] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2ee89b-7640-4ae6-bf4d-cd337e31a2f1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.660104] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ce4aed-02d9-4d10-a02d-326fde6c89f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.688914] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] volume-1d559b9c-feac-4939-85bd-15d6203183b4/volume-1d559b9c-feac-4939-85bd-15d6203183b4.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.689358] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83126c48-e2aa-4a48-a9f9-c95bfe68bd7b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.709690] env[61936]: DEBUG oslo_vmware.api [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1018.709690] env[61936]: value = "task-1253373" [ 1018.709690] env[61936]: _type = "Task" [ 1018.709690] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.718851] env[61936]: DEBUG oslo_vmware.api [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253373, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.757515] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.760260] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.506s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.762157] env[61936]: INFO nova.compute.claims [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1018.788727] env[61936]: INFO nova.scheduler.client.report [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted allocations for instance b780470b-57d0-43e5-be52-539c78cada6b [ 1018.907101] env[61936]: DEBUG nova.objects.instance [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'pci_requests' on Instance uuid 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.923744] env[61936]: DEBUG nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1018.938097] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527f2026-0c74-5e37-fd69-56edf94edd72/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1018.939580] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cead75f-084d-46b1-a581-e199b984c60d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.949998] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527f2026-0c74-5e37-fd69-56edf94edd72/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1018.950220] env[61936]: ERROR oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527f2026-0c74-5e37-fd69-56edf94edd72/disk-0.vmdk due to incomplete transfer. [ 1018.952369] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1018.952608] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.952771] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1018.952944] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.953107] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1018.953260] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1018.953469] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1018.953666] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1018.954653] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1018.954653] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1018.954653] env[61936]: DEBUG nova.virt.hardware [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1018.954653] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-23f674d0-02aa-41d4-944d-bb2010427cba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.956426] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339c16df-1ea1-47c0-b0c5-8edec991ca8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.964417] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd83e95d-17bd-4511-981d-2cf9cf5ebd21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.969058] env[61936]: DEBUG oslo_vmware.rw_handles [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527f2026-0c74-5e37-fd69-56edf94edd72/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1018.969284] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Uploaded image baa3a86e-79f6-4c48-b14b-11d628f6dec2 to the Glance image server {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1018.971680] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Destroying the VM {{(pid=61936) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1018.972444] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f3aae660-cb27-4cb3-8042-1257474fa933 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.986064] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1018.986064] env[61936]: value = "task-1253374" [ 1018.986064] env[61936]: _type = "Task" [ 1018.986064] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.996696] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253374, 'name': Destroy_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.012535] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253372, 'name': PowerOffVM_Task, 'duration_secs': 0.189954} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.012797] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.012981] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance '755c5fea-e3d2-4cb3-b717-5f862401d7e1' progress to 17 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1019.221146] env[61936]: DEBUG oslo_vmware.api [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253373, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.296145] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49adde3-d437-4c1e-b68a-a9a4f500591a tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "b780470b-57d0-43e5-be52-539c78cada6b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.965s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.305089] env[61936]: DEBUG nova.network.neutron [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updated VIF entry in instance network info cache for port b556882a-8fa4-4ea2-871b-6784ef87c88c. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.305451] env[61936]: DEBUG nova.network.neutron [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updating instance_info_cache with network_info: [{"id": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "address": "fa:16:3e:44:81:ac", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb556882a-8f", "ovs_interfaceid": "b556882a-8fa4-4ea2-871b-6784ef87c88c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.410346] env[61936]: DEBUG nova.objects.base [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Object Instance<0137305a-dace-4eda-9d90-7233c001176f> lazy-loaded attributes: flavor,pci_requests {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1019.410606] env[61936]: DEBUG nova.network.neutron [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1019.488547] env[61936]: DEBUG nova.policy [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1019.498630] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253374, 'name': Destroy_Task, 'duration_secs': 0.395079} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.498888] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Destroyed the VM [ 1019.499177] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Deleting Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1019.499419] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d301bb74-9eb8-42d4-8197-1cefddbcc5af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.505618] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1019.505618] env[61936]: value = "task-1253375" [ 1019.505618] env[61936]: _type = "Task" [ 1019.505618] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.515842] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253375, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.519694] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1019.520008] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.520259] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1019.520507] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.520716] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1019.520932] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1019.521247] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1019.521483] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1019.521764] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1019.522013] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1019.522277] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1019.530381] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ca6c987-e208-4e30-b357-1fee8e6287a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.552383] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1019.552383] env[61936]: value = "task-1253376" [ 1019.552383] env[61936]: _type = "Task" [ 1019.552383] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.566126] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253376, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.721053] env[61936]: DEBUG oslo_vmware.api [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253373, 'name': ReconfigVM_Task, 'duration_secs': 0.923708} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.721393] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Reconfigured VM instance instance-00000056 to attach disk [datastore1] volume-1d559b9c-feac-4939-85bd-15d6203183b4/volume-1d559b9c-feac-4939-85bd-15d6203183b4.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.726345] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca5fc5e2-4d70-4024-908a-5c4c84042d14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.743571] env[61936]: DEBUG oslo_vmware.api [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1019.743571] env[61936]: value = "task-1253377" [ 1019.743571] env[61936]: _type = "Task" [ 1019.743571] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.752266] env[61936]: DEBUG oslo_vmware.api [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253377, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.807937] env[61936]: DEBUG oslo_concurrency.lockutils [req-cf0bf095-bc56-4405-8bb3-1c55563f2818 req-4886ca28-0387-462a-8413-678fbb89c411 service nova] Releasing lock "refresh_cache-18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.013010] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a9317f-3d68-4c48-ab7a-54c970b6e23a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.019458] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253375, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.024503] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad13d9b3-3163-429c-8b15-f487af41dbae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.061584] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87283bfb-4333-4103-9b82-0713c94269b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.072377] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253376, 'name': ReconfigVM_Task, 'duration_secs': 0.242044} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.072757] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance '755c5fea-e3d2-4cb3-b717-5f862401d7e1' progress to 33 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1020.080021] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b88e5b3-08aa-48bd-91b7-620cab97f52b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.092524] env[61936]: DEBUG nova.compute.provider_tree [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.253858] env[61936]: DEBUG oslo_vmware.api [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253377, 'name': ReconfigVM_Task, 'duration_secs': 0.161992} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.254285] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270021', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'name': 'volume-1d559b9c-feac-4939-85bd-15d6203183b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8', 'attached_at': '', 'detached_at': '', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'serial': '1d559b9c-feac-4939-85bd-15d6203183b4'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1020.277692] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.389161] env[61936]: DEBUG nova.network.neutron [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Successfully created port: 3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.517524] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253375, 'name': RemoveSnapshot_Task, 'duration_secs': 0.992334} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.517946] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Deleted Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1020.518276] env[61936]: DEBUG nova.compute.manager [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1020.519143] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56c76d1-0fd4-4485-b61e-30dddf81777c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.585122] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1020.585453] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.585673] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1020.585931] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.586174] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1020.586467] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1020.586788] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1020.587153] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1020.587302] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1020.587450] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1020.587686] env[61936]: DEBUG nova.virt.hardware [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1020.596982] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1020.596982] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfcc30ce-a813-4a4f-aab1-4c29c32ba6a5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.608492] env[61936]: DEBUG nova.scheduler.client.report [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1020.617785] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1020.617785] env[61936]: value = "task-1253378" [ 1020.617785] env[61936]: _type = "Task" [ 1020.617785] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.626664] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253378, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.697525] env[61936]: DEBUG nova.compute.manager [req-8dd79700-e95b-42bb-8fca-bdc6f3d0e785 req-bfaa193b-48de-4735-b3e1-3751fb5195e9 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Received event network-vif-plugged-a424508f-87c8-48b0-b5de-4a4916cddc3f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1020.697817] env[61936]: DEBUG oslo_concurrency.lockutils [req-8dd79700-e95b-42bb-8fca-bdc6f3d0e785 req-bfaa193b-48de-4735-b3e1-3751fb5195e9 service nova] Acquiring lock "1c391c45-a041-422d-9a44-a29306f99a6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.698251] env[61936]: DEBUG oslo_concurrency.lockutils [req-8dd79700-e95b-42bb-8fca-bdc6f3d0e785 req-bfaa193b-48de-4735-b3e1-3751fb5195e9 service nova] Lock "1c391c45-a041-422d-9a44-a29306f99a6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.698588] env[61936]: DEBUG oslo_concurrency.lockutils [req-8dd79700-e95b-42bb-8fca-bdc6f3d0e785 req-bfaa193b-48de-4735-b3e1-3751fb5195e9 service nova] Lock "1c391c45-a041-422d-9a44-a29306f99a6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.698890] env[61936]: DEBUG nova.compute.manager [req-8dd79700-e95b-42bb-8fca-bdc6f3d0e785 req-bfaa193b-48de-4735-b3e1-3751fb5195e9 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] No waiting events found dispatching network-vif-plugged-a424508f-87c8-48b0-b5de-4a4916cddc3f {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1020.699453] env[61936]: WARNING nova.compute.manager [req-8dd79700-e95b-42bb-8fca-bdc6f3d0e785 req-bfaa193b-48de-4735-b3e1-3751fb5195e9 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Received unexpected event network-vif-plugged-a424508f-87c8-48b0-b5de-4a4916cddc3f for instance with vm_state building and task_state spawning. [ 1020.711680] env[61936]: DEBUG nova.network.neutron [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Successfully updated port: a424508f-87c8-48b0-b5de-4a4916cddc3f {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.780895] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.781078] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updated the network info_cache for instance {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10397}} [ 1020.781504] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1020.781765] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1020.781956] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1020.782164] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1020.782386] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1020.782535] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1020.782663] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 1020.782851] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.034967] env[61936]: INFO nova.compute.manager [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Shelve offloading [ 1021.106812] env[61936]: DEBUG nova.compute.manager [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Received event network-changed-a424508f-87c8-48b0-b5de-4a4916cddc3f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1021.107192] env[61936]: DEBUG nova.compute.manager [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Refreshing instance network info cache due to event network-changed-a424508f-87c8-48b0-b5de-4a4916cddc3f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1021.107552] env[61936]: DEBUG oslo_concurrency.lockutils [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] Acquiring lock "refresh_cache-1c391c45-a041-422d-9a44-a29306f99a6d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.107803] env[61936]: DEBUG oslo_concurrency.lockutils [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] Acquired lock "refresh_cache-1c391c45-a041-422d-9a44-a29306f99a6d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.108114] env[61936]: DEBUG nova.network.neutron [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Refreshing network info cache for port a424508f-87c8-48b0-b5de-4a4916cddc3f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.113132] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.113831] env[61936]: DEBUG nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1021.118153] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.894s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.118923] env[61936]: DEBUG nova.objects.instance [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lazy-loading 'resources' on Instance uuid 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.134663] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253378, 'name': ReconfigVM_Task, 'duration_secs': 0.341339} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.135011] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1021.136239] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982a5fe4-5f08-481e-9dfb-efe3f156c099 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.876632] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "refresh_cache-1c391c45-a041-422d-9a44-a29306f99a6d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.878843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.879529] env[61936]: DEBUG nova.objects.instance [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.880959] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.884032] env[61936]: DEBUG nova.compute.utils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1021.894528] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 755c5fea-e3d2-4cb3-b717-5f862401d7e1/755c5fea-e3d2-4cb3-b717-5f862401d7e1.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.896724] env[61936]: DEBUG nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1021.898605] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0bdf299-d895-499d-a6a1-7830925988db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.903023] env[61936]: DEBUG nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1021.903023] env[61936]: DEBUG nova.network.neutron [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1021.903023] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a9d7f7d-bdec-4722-812c-9b097140f20d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.915918] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51f09f1-b506-4db3-8241-2acda738ca5d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.920160] env[61936]: DEBUG nova.network.neutron [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Successfully updated port: 3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.927537] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b6584769-087c-4b57-8f1f-4ab59344e57d tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Suspending the VM {{(pid=61936) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1021.930537] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4759920a-2b9e-429d-8333-f0a9265638c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.932975] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1021.932975] env[61936]: value = "task-1253379" [ 1021.932975] env[61936]: _type = "Task" [ 1021.932975] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.934711] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1021.934711] env[61936]: value = "task-1253380" [ 1021.934711] env[61936]: _type = "Task" [ 1021.934711] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.942907] env[61936]: DEBUG oslo_vmware.api [None req-b6584769-087c-4b57-8f1f-4ab59344e57d tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1021.942907] env[61936]: value = "task-1253381" [ 1021.942907] env[61936]: _type = "Task" [ 1021.942907] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.954743] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253380, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.955586] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1021.955720] env[61936]: DEBUG nova.compute.manager [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1021.959763] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448e336d-15a9-4b7c-8751-ce7cfd90d799 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.967027] env[61936]: DEBUG oslo_vmware.api [None req-b6584769-087c-4b57-8f1f-4ab59344e57d tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253381, 'name': SuspendVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.969151] env[61936]: DEBUG nova.network.neutron [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1021.972197] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.972361] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.972581] env[61936]: DEBUG nova.network.neutron [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.999138] env[61936]: DEBUG nova.policy [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcee2be71ae044338df4422a3c575ca4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e81b119251964bdb8e5a0ea84b29f2a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1022.131242] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0d5d48-ff7a-47c3-a118-ba5d33660ede {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.134940] env[61936]: DEBUG nova.network.neutron [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.141188] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b56f839-4cd3-4070-b934-ab548372f903 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.174712] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6879ee6d-eb49-42c6-9d6f-7c4493f01d21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.182922] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da74438b-3a41-480d-bf19-50d3281141b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.196938] env[61936]: DEBUG nova.compute.provider_tree [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.398547] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f6c94dc6-a103-47b0-b826-ecf905e064b4 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.385s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.423720] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.424008] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.424274] env[61936]: DEBUG nova.network.neutron [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1022.465022] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253380, 'name': ReconfigVM_Task, 'duration_secs': 0.277304} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.465325] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 755c5fea-e3d2-4cb3-b717-5f862401d7e1/755c5fea-e3d2-4cb3-b717-5f862401d7e1.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.465398] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance '755c5fea-e3d2-4cb3-b717-5f862401d7e1' progress to 50 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1022.469966] env[61936]: DEBUG oslo_vmware.api [None req-b6584769-087c-4b57-8f1f-4ab59344e57d tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253381, 'name': SuspendVM_Task} progress is 54%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.479258] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.479444] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.479626] env[61936]: DEBUG nova.compute.manager [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1022.480593] env[61936]: DEBUG nova.network.neutron [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Successfully created port: 69b0f1f9-2468-4011-bab2-0006651f7bba {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1022.483769] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b1dd52-920f-4b01-9fe8-fb9a6aafada5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.490631] env[61936]: DEBUG nova.compute.manager [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61936) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1022.491199] env[61936]: DEBUG nova.objects.instance [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.639247] env[61936]: DEBUG oslo_concurrency.lockutils [req-9f5e3154-7140-4623-912c-db733abeaca0 req-46c66135-c836-423b-8a91-37d706f82810 service nova] Releasing lock "refresh_cache-1c391c45-a041-422d-9a44-a29306f99a6d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.639247] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "refresh_cache-1c391c45-a041-422d-9a44-a29306f99a6d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.639247] env[61936]: DEBUG nova.network.neutron [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1022.702025] env[61936]: DEBUG nova.scheduler.client.report [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1022.926970] env[61936]: DEBUG nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1022.955578] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1022.956576] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.956822] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1022.957096] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.957326] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1022.957550] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1022.958366] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1022.958366] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1022.958366] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1022.958366] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1022.958664] env[61936]: DEBUG nova.virt.hardware [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1022.959609] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5c9366-c6f7-43a6-b756-7534957cea16 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.972658] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663e1004-6488-49e9-97a6-9b8ba44b8399 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.979437] env[61936]: DEBUG oslo_vmware.api [None req-b6584769-087c-4b57-8f1f-4ab59344e57d tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253381, 'name': SuspendVM_Task, 'duration_secs': 0.956108} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.979437] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fadafda-ef09-4f96-811f-11cddc2f21ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.982110] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b6584769-087c-4b57-8f1f-4ab59344e57d tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Suspended the VM {{(pid=61936) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1022.984587] env[61936]: DEBUG nova.compute.manager [None req-b6584769-087c-4b57-8f1f-4ab59344e57d tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1022.986263] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a39c02f-95d6-4d0a-b9bb-f4d886882cde {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.062587] env[61936]: WARNING nova.network.neutron [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] 03166185-d5e0-477b-892e-20bea284c8db already exists in list: networks containing: ['03166185-d5e0-477b-892e-20bea284c8db']. ignoring it [ 1023.062587] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb27d06d-fcc8-46f4-8d8e-d33cb9889bf9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.062587] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance '755c5fea-e3d2-4cb3-b717-5f862401d7e1' progress to 67 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1023.092015] env[61936]: DEBUG nova.network.neutron [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [{"id": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "address": "fa:16:3e:65:44:c9", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd169ca50-5a", "ovs_interfaceid": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.199342] env[61936]: DEBUG nova.network.neutron [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1023.207245] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.089s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.209797] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.331s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.210034] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.210606] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1023.211218] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49eaa82-d832-414c-9dd2-0e0caa76cb74 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.220326] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2f1132-38cc-479c-b803-bf51bcff7f55 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.237386] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9423c1-86d2-429a-950c-b5fdee256196 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.247373] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e570bcfa-8bfb-416f-a7b1-61c6dcb83b0c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.253167] env[61936]: INFO nova.scheduler.client.report [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted allocations for instance 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16 [ 1023.286214] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180003MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1023.286412] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.286601] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.378903] env[61936]: DEBUG nova.compute.manager [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-vif-plugged-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1023.379142] env[61936]: DEBUG oslo_concurrency.lockutils [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.379422] env[61936]: DEBUG oslo_concurrency.lockutils [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.379639] env[61936]: DEBUG oslo_concurrency.lockutils [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.379874] env[61936]: DEBUG nova.compute.manager [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] No waiting events found dispatching network-vif-plugged-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1023.379967] env[61936]: WARNING nova.compute.manager [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received unexpected event network-vif-plugged-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 for instance with vm_state active and task_state None. [ 1023.380190] env[61936]: DEBUG nova.compute.manager [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-changed-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1023.380290] env[61936]: DEBUG nova.compute.manager [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing instance network info cache due to event network-changed-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1023.380502] env[61936]: DEBUG oslo_concurrency.lockutils [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.526407] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1023.526658] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05b15ee6-3cbe-47b6-8a3b-bbaf9451a00a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.533387] env[61936]: DEBUG nova.network.neutron [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Updating instance_info_cache with network_info: [{"id": "a424508f-87c8-48b0-b5de-4a4916cddc3f", "address": "fa:16:3e:31:f1:db", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa424508f-87", "ovs_interfaceid": "a424508f-87c8-48b0-b5de-4a4916cddc3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.535844] env[61936]: DEBUG oslo_vmware.api [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1023.535844] env[61936]: value = "task-1253382" [ 1023.535844] env[61936]: _type = "Task" [ 1023.535844] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.547309] env[61936]: DEBUG oslo_vmware.api [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.595904] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.652927] env[61936]: DEBUG nova.network.neutron [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Port 7775677e-cf35-458b-8c94-da7dd8dc652a binding to destination host cpu-1 is already ACTIVE {{(pid=61936) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1023.707457] env[61936]: DEBUG nova.network.neutron [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "address": "fa:16:3e:fe:29:21", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ccd46c0-94", "ovs_interfaceid": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.798187] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2002c868-1129-4b1b-8454-3610c2f8731b tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "8d6c05f2-4bac-41ce-a9f0-4df3d86eba16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.398s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.004888] env[61936]: DEBUG nova.compute.manager [req-b10fde0b-aa92-41e2-b096-01dc718c1cfe req-c8f01e09-8042-4c4e-8ef2-a4dca9fc4f22 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-vif-unplugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1024.005155] env[61936]: DEBUG oslo_concurrency.lockutils [req-b10fde0b-aa92-41e2-b096-01dc718c1cfe req-c8f01e09-8042-4c4e-8ef2-a4dca9fc4f22 service nova] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.005358] env[61936]: DEBUG oslo_concurrency.lockutils [req-b10fde0b-aa92-41e2-b096-01dc718c1cfe req-c8f01e09-8042-4c4e-8ef2-a4dca9fc4f22 service nova] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.005529] env[61936]: DEBUG oslo_concurrency.lockutils [req-b10fde0b-aa92-41e2-b096-01dc718c1cfe req-c8f01e09-8042-4c4e-8ef2-a4dca9fc4f22 service nova] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.005868] env[61936]: DEBUG nova.compute.manager [req-b10fde0b-aa92-41e2-b096-01dc718c1cfe req-c8f01e09-8042-4c4e-8ef2-a4dca9fc4f22 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] No waiting events found dispatching network-vif-unplugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1024.006058] env[61936]: WARNING nova.compute.manager [req-b10fde0b-aa92-41e2-b096-01dc718c1cfe req-c8f01e09-8042-4c4e-8ef2-a4dca9fc4f22 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received unexpected event network-vif-unplugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 for instance with vm_state shelved and task_state shelving_offloading. [ 1024.037543] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "refresh_cache-1c391c45-a041-422d-9a44-a29306f99a6d" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.037889] env[61936]: DEBUG nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Instance network_info: |[{"id": "a424508f-87c8-48b0-b5de-4a4916cddc3f", "address": "fa:16:3e:31:f1:db", "network": {"id": "ebddf4a3-56c8-450b-8c5b-17c358b0d8b2", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1838459928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1b7d01ffac74c6fa2306b4839041fdd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa424508f-87", "ovs_interfaceid": "a424508f-87c8-48b0-b5de-4a4916cddc3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1024.038786] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:f1:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a424508f-87c8-48b0-b5de-4a4916cddc3f', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.051062] env[61936]: DEBUG oslo.service.loopingcall [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.056164] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1024.056476] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7807ab5-07e0-4b18-b940-55c918bbb557 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.083433] env[61936]: DEBUG oslo_vmware.api [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253382, 'name': PowerOffVM_Task, 'duration_secs': 0.28539} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.084747] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1024.084953] env[61936]: DEBUG nova.compute.manager [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1024.085227] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.085227] env[61936]: value = "task-1253383" [ 1024.085227] env[61936]: _type = "Task" [ 1024.085227] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.085922] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92edbd3-b112-4dc8-a085-b7053806101e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.106050] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253383, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.177661] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.178608] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec293f97-c512-41b8-b774-4737df6deae2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.186538] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1024.186802] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c97e5846-7dbd-47cc-93cb-cb6e1040e708 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.209783] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.210519] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.210681] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.211248] env[61936]: DEBUG oslo_concurrency.lockutils [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.211814] env[61936]: DEBUG nova.network.neutron [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing network info cache for port 3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.213722] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bad5daa-a700-4f9b-b5ea-38f978515dfb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.232056] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1024.232287] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.232448] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1024.232635] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.232817] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1024.232953] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1024.233143] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1024.233331] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1024.233543] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1024.233713] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1024.233889] env[61936]: DEBUG nova.virt.hardware [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1024.240198] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfiguring VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1024.240836] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cfb6f97-a592-4f8d-a5cd-e1cb53109bb2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.258806] env[61936]: DEBUG oslo_vmware.api [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1024.258806] env[61936]: value = "task-1253385" [ 1024.258806] env[61936]: _type = "Task" [ 1024.258806] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.267264] env[61936]: DEBUG oslo_vmware.api [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253385, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.276862] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1024.277178] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1024.277440] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleting the datastore file [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.277764] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3819fed3-b01b-459a-af7f-4187fd7268a6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.284826] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1024.284826] env[61936]: value = "task-1253386" [ 1024.284826] env[61936]: _type = "Task" [ 1024.284826] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.294732] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253386, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.302127] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Applying migration context for instance 755c5fea-e3d2-4cb3-b717-5f862401d7e1 as it has an incoming, in-progress migration 2d0e61d3-548f-4a27-a434-95d93432350b. Migration status is post-migrating {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1024.304325] env[61936]: INFO nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating resource usage from migration 2d0e61d3-548f-4a27-a434-95d93432350b [ 1024.329996] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fa845f8e-957c-4c0b-a647-190c32989dcd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.330157] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f3f2a0a8-cfe3-462f-a940-05580d5da32d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.330289] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fbde880a-b47c-49e8-b84d-0efee21d62c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.330454] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.330601] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 0137305a-dace-4eda-9d90-7233c001176f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.330764] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 484e5d41-29f4-4845-9633-157c03766978 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.330884] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 5526cbd5-b1ad-453b-8401-eee7aa356606 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.330996] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e6930eaf-bb63-4553-8cd2-f2eb62fca510 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.331128] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.331243] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Migration 2d0e61d3-548f-4a27-a434-95d93432350b is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1024.332759] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 755c5fea-e3d2-4cb3-b717-5f862401d7e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.333326] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 1c391c45-a041-422d-9a44-a29306f99a6d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.333326] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 53605e96-a4d1-4a7b-8fef-1f1878b099c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.333423] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1024.333512] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1024.514977] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76f9f56-53a8-48ff-83af-130e11ac2976 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.526022] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898f41ff-b6c9-4e8e-a9ab-e08984237c23 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.556196] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59c9b6b-3622-4683-913f-193454ada10c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.564485] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f661443-d508-45a9-bc30-e68a1d97a8f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.579783] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.602043] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253383, 'name': CreateVM_Task, 'duration_secs': 0.383382} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.602043] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1024.602043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.602043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.602043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.602043] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9901709-3c30-4f4f-8b5a-93fe0d1d2470 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.603025] env[61936]: DEBUG nova.network.neutron [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Successfully updated port: 69b0f1f9-2468-4011-bab2-0006651f7bba {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1024.608019] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1024.608019] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522520cf-fd88-4ff4-b9f9-f6e4b6709481" [ 1024.608019] env[61936]: _type = "Task" [ 1024.608019] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.611444] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ce1aaddb-c7da-4f2d-ae58-664b96a52091 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.132s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.617068] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]522520cf-fd88-4ff4-b9f9-f6e4b6709481, 'name': SearchDatastore_Task, 'duration_secs': 0.009014} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.617068] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.617240] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.617522] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.617682] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.617881] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.618175] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-241fefa6-6782-414e-8bb0-abe9f8235165 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.626879] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.627095] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1024.628224] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e65f995b-9d4b-46f5-a195-2e9cf37396f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.633606] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1024.633606] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525f1aeb-1488-31a8-c919-71b3b902b40f" [ 1024.633606] env[61936]: _type = "Task" [ 1024.633606] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.641572] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525f1aeb-1488-31a8-c919-71b3b902b40f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.673791] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.674074] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.674261] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.768934] env[61936]: DEBUG oslo_vmware.api [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253385, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.793556] env[61936]: DEBUG oslo_vmware.api [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253386, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300922} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.793746] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.793938] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.794167] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.816527] env[61936]: INFO nova.scheduler.client.report [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted allocations for instance fbde880a-b47c-49e8-b84d-0efee21d62c0 [ 1024.964963] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.965253] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.023977] env[61936]: DEBUG nova.network.neutron [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updated VIF entry in instance network info cache for port 3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.024480] env[61936]: DEBUG nova.network.neutron [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "address": "fa:16:3e:fe:29:21", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ccd46c0-94", "ovs_interfaceid": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.082783] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1025.107140] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-53605e96-a4d1-4a7b-8fef-1f1878b099c3" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.107281] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-53605e96-a4d1-4a7b-8fef-1f1878b099c3" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.107435] env[61936]: DEBUG nova.network.neutron [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.144718] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]525f1aeb-1488-31a8-c919-71b3b902b40f, 'name': SearchDatastore_Task, 'duration_secs': 0.007402} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.145921] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76d35aa3-b22b-4d00-ba0a-117f81e6920a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.150734] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1025.150734] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52badd21-45dc-63a7-ebfd-bffd786f42f8" [ 1025.150734] env[61936]: _type = "Task" [ 1025.150734] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.158307] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52badd21-45dc-63a7-ebfd-bffd786f42f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.229858] env[61936]: INFO nova.compute.manager [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Resuming [ 1025.230530] env[61936]: DEBUG nova.objects.instance [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lazy-loading 'flavor' on Instance uuid 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.269180] env[61936]: DEBUG oslo_vmware.api [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253385, 'name': ReconfigVM_Task, 'duration_secs': 0.755248} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.269708] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.269922] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfigured VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1025.327197] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.403396] env[61936]: DEBUG nova.compute.manager [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Received event network-vif-plugged-69b0f1f9-2468-4011-bab2-0006651f7bba {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1025.403621] env[61936]: DEBUG oslo_concurrency.lockutils [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] Acquiring lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.403824] env[61936]: DEBUG oslo_concurrency.lockutils [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.403991] env[61936]: DEBUG oslo_concurrency.lockutils [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.404177] env[61936]: DEBUG nova.compute.manager [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] No waiting events found dispatching network-vif-plugged-69b0f1f9-2468-4011-bab2-0006651f7bba {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1025.404359] env[61936]: WARNING nova.compute.manager [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Received unexpected event network-vif-plugged-69b0f1f9-2468-4011-bab2-0006651f7bba for instance with vm_state building and task_state spawning. [ 1025.404491] env[61936]: DEBUG nova.compute.manager [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Received event network-changed-69b0f1f9-2468-4011-bab2-0006651f7bba {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1025.404646] env[61936]: DEBUG nova.compute.manager [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Refreshing instance network info cache due to event network-changed-69b0f1f9-2468-4011-bab2-0006651f7bba. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1025.404809] env[61936]: DEBUG oslo_concurrency.lockutils [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] Acquiring lock "refresh_cache-53605e96-a4d1-4a7b-8fef-1f1878b099c3" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.467625] env[61936]: DEBUG nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1025.527716] env[61936]: DEBUG oslo_concurrency.lockutils [req-a082d600-b4f8-4787-97f6-7842cd069d50 req-ace2fd7f-94a7-4367-b394-81207597e387 service nova] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.588553] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1025.588813] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.302s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.589072] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.262s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.589345] env[61936]: DEBUG nova.objects.instance [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'resources' on Instance uuid fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.645059] env[61936]: DEBUG nova.network.neutron [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1025.660561] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52badd21-45dc-63a7-ebfd-bffd786f42f8, 'name': SearchDatastore_Task, 'duration_secs': 0.023697} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.660872] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.661147] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1c391c45-a041-422d-9a44-a29306f99a6d/1c391c45-a041-422d-9a44-a29306f99a6d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1025.661404] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89e4e835-004e-4032-acd8-d0213b6dabbf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.672615] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1025.672615] env[61936]: value = "task-1253387" [ 1025.672615] env[61936]: _type = "Task" [ 1025.672615] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.685821] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.736223] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.736424] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.736603] env[61936]: DEBUG nova.network.neutron [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.751546] env[61936]: DEBUG nova.objects.instance [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.774796] env[61936]: DEBUG oslo_concurrency.lockutils [None req-919191bc-8984-4392-8da5-fa6ee88d7fbe tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.502s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.892165] env[61936]: DEBUG nova.network.neutron [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Updating instance_info_cache with network_info: [{"id": "69b0f1f9-2468-4011-bab2-0006651f7bba", "address": "fa:16:3e:95:ae:7e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69b0f1f9-24", "ovs_interfaceid": "69b0f1f9-2468-4011-bab2-0006651f7bba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.029310] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.092891] env[61936]: DEBUG nova.objects.instance [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'numa_topology' on Instance uuid fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.183350] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253387, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.256593] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.256826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.257052] env[61936]: DEBUG nova.network.neutron [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.257304] env[61936]: DEBUG nova.objects.instance [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'info_cache' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.395129] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-53605e96-a4d1-4a7b-8fef-1f1878b099c3" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.395733] env[61936]: DEBUG nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Instance network_info: |[{"id": "69b0f1f9-2468-4011-bab2-0006651f7bba", "address": "fa:16:3e:95:ae:7e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69b0f1f9-24", "ovs_interfaceid": "69b0f1f9-2468-4011-bab2-0006651f7bba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1026.396095] env[61936]: DEBUG oslo_concurrency.lockutils [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] Acquired lock "refresh_cache-53605e96-a4d1-4a7b-8fef-1f1878b099c3" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.396396] env[61936]: DEBUG nova.network.neutron [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Refreshing network info cache for port 69b0f1f9-2468-4011-bab2-0006651f7bba {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1026.398174] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:ae:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69b0f1f9-2468-4011-bab2-0006651f7bba', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.409064] env[61936]: DEBUG oslo.service.loopingcall [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.412275] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1026.412786] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8906b7ff-ffbe-4a10-9c3e-4cd9375ccca5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.432353] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.432353] env[61936]: value = "task-1253388" [ 1026.432353] env[61936]: _type = "Task" [ 1026.432353] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.440304] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253388, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.482662] env[61936]: DEBUG nova.compute.manager [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1026.482921] env[61936]: DEBUG nova.compute.manager [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing instance network info cache due to event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1026.483679] env[61936]: DEBUG oslo_concurrency.lockutils [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] Acquiring lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.483679] env[61936]: DEBUG oslo_concurrency.lockutils [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] Acquired lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.483679] env[61936]: DEBUG nova.network.neutron [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing network info cache for port d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1026.595640] env[61936]: DEBUG nova.objects.base [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1026.686405] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253387, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.771614} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.687105] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 1c391c45-a041-422d-9a44-a29306f99a6d/1c391c45-a041-422d-9a44-a29306f99a6d.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1026.687389] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.687701] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-475f724f-ea96-4f28-ab01-bb806ee2bc38 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.694100] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1026.694100] env[61936]: value = "task-1253389" [ 1026.694100] env[61936]: _type = "Task" [ 1026.694100] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.703998] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253389, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.752476] env[61936]: DEBUG oslo_concurrency.lockutils [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.752672] env[61936]: DEBUG oslo_concurrency.lockutils [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquired lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.752860] env[61936]: DEBUG nova.network.neutron [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.761063] env[61936]: DEBUG nova.objects.base [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Object Instance<0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8> lazy-loaded attributes: flavor,info_cache {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1026.788093] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35898c5-1c1e-4f35-9d6d-57929a10e3e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.800111] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659301c5-e3d4-42bd-b754-32c19899ece3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.835786] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3e21a3-f097-48d4-8b2b-320481980600 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.844113] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578c6656-2bdb-458c-a366-c2446e256ba3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.862011] env[61936]: DEBUG nova.compute.provider_tree [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.913951] env[61936]: DEBUG nova.network.neutron [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance_info_cache with network_info: [{"id": "7775677e-cf35-458b-8c94-da7dd8dc652a", "address": "fa:16:3e:f7:93:44", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7775677e-cf", "ovs_interfaceid": "7775677e-cf35-458b-8c94-da7dd8dc652a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.944734] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253388, 'name': CreateVM_Task, 'duration_secs': 0.449041} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.947009] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1026.947009] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.947009] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.947009] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1026.947328] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-394a02c4-dbe7-44b4-8465-a0f36e0fb16b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.953498] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1026.953498] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52556e07-82ca-7cc6-3849-bd0956cca7ae" [ 1026.953498] env[61936]: _type = "Task" [ 1026.953498] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.963238] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52556e07-82ca-7cc6-3849-bd0956cca7ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.203975] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253389, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069249} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.208045] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.208045] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00c12c8-b072-4e77-93b4-e913aabed075 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.231464] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 1c391c45-a041-422d-9a44-a29306f99a6d/1c391c45-a041-422d-9a44-a29306f99a6d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.232664] env[61936]: DEBUG nova.network.neutron [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Updated VIF entry in instance network info cache for port 69b0f1f9-2468-4011-bab2-0006651f7bba. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1027.233147] env[61936]: DEBUG nova.network.neutron [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Updating instance_info_cache with network_info: [{"id": "69b0f1f9-2468-4011-bab2-0006651f7bba", "address": "fa:16:3e:95:ae:7e", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69b0f1f9-24", "ovs_interfaceid": "69b0f1f9-2468-4011-bab2-0006651f7bba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.234463] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f13ab056-25fd-40e7-94d8-016d3f3a61cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.262236] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1027.262236] env[61936]: value = "task-1253390" [ 1027.262236] env[61936]: _type = "Task" [ 1027.262236] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.273449] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.370437] env[61936]: DEBUG nova.scheduler.client.report [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1027.377130] env[61936]: DEBUG nova.network.neutron [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updated VIF entry in instance network info cache for port d169ca50-5a14-4a65-9da0-ea79c80f75c8. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1027.377601] env[61936]: DEBUG nova.network.neutron [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [{"id": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "address": "fa:16:3e:65:44:c9", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd169ca50-5a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.417472] env[61936]: DEBUG oslo_concurrency.lockutils [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.468368] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52556e07-82ca-7cc6-3849-bd0956cca7ae, 'name': SearchDatastore_Task, 'duration_secs': 0.009668} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.471281] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.471623] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.471937] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.472152] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.473039] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.473039] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e620bf58-8a64-49dc-9b56-bb035c9e5eb0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.484057] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.484057] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1027.484734] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d833fc4-a44a-4cce-9df3-89c1fe7d47e7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.491897] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1027.491897] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5256416f-b8b0-b47f-c743-460dadbf8ca2" [ 1027.491897] env[61936]: _type = "Task" [ 1027.491897] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.500796] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5256416f-b8b0-b47f-c743-460dadbf8ca2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.510636] env[61936]: DEBUG nova.network.neutron [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [{"id": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "address": "fa:16:3e:84:fd:4f", "network": {"id": "e8052ea1-3ae4-4cbc-9f6c-db7ee942b296", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-576975673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c334e9711bc64304ab017eb6ab3e2d23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cdd9a15-01", "ovs_interfaceid": "1cdd9a15-01e7-40cf-80fc-24bea0b642d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.590412] env[61936]: DEBUG nova.network.neutron [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating instance_info_cache with network_info: [{"id": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "address": "fa:16:3e:ce:3d:0e", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746d118b-a8", "ovs_interfaceid": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.751681] env[61936]: DEBUG oslo_concurrency.lockutils [req-a860e4d6-5d47-4001-baec-f272681e35d8 req-0142228c-a988-4001-9936-560e62c8d2d3 service nova] Releasing lock "refresh_cache-53605e96-a4d1-4a7b-8fef-1f1878b099c3" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.772105] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253390, 'name': ReconfigVM_Task, 'duration_secs': 0.297182} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.772392] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 1c391c45-a041-422d-9a44-a29306f99a6d/1c391c45-a041-422d-9a44-a29306f99a6d.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.773023] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5ad9020-292c-4afc-9630-a831c3b5cd6e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.779712] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1027.779712] env[61936]: value = "task-1253391" [ 1027.779712] env[61936]: _type = "Task" [ 1027.779712] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.802727] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253391, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.819772] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.880245] env[61936]: DEBUG oslo_concurrency.lockutils [req-ab1ba5f2-1c44-425c-a417-411d25bffb6f req-ce70811d-8124-4373-8023-dee3be133e7c service nova] Releasing lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.882309] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.293s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.884800] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.856s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.886348] env[61936]: INFO nova.compute.claims [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1027.945251] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e541e32e-ab80-4de6-9e98-718e45e81ccb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.968622] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e9eb02-f15b-413f-bbb7-7fdef333282f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.971716] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-0137305a-dace-4eda-9d90-7233c001176f-fa637c42-69ec-480c-bc7b-08a8a723660b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.971951] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-fa637c42-69ec-480c-bc7b-08a8a723660b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.972317] env[61936]: DEBUG nova.objects.instance [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'flavor' on Instance uuid 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.979157] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance '755c5fea-e3d2-4cb3-b717-5f862401d7e1' progress to 83 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1028.002126] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5256416f-b8b0-b47f-c743-460dadbf8ca2, 'name': SearchDatastore_Task, 'duration_secs': 0.008776} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.004261] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-080fe244-f649-4fb4-96db-1e1e181d5cd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.010534] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1028.010534] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52851da9-af44-f01a-d3c5-ab08e3b5c297" [ 1028.010534] env[61936]: _type = "Task" [ 1028.010534] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.013999] env[61936]: DEBUG oslo_concurrency.lockutils [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Releasing lock "refresh_cache-5526cbd5-b1ad-453b-8401-eee7aa356606" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.014780] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9533da83-1984-4754-b2c1-748fa8f61f2b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.022370] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Resuming the VM {{(pid=61936) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1028.026043] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b37932f-9f78-4e03-9548-734d45d9e327 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.027676] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52851da9-af44-f01a-d3c5-ab08e3b5c297, 'name': SearchDatastore_Task, 'duration_secs': 0.010496} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.027916] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.028172] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 53605e96-a4d1-4a7b-8fef-1f1878b099c3/53605e96-a4d1-4a7b-8fef-1f1878b099c3.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.028721] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a714eda-bc2d-4f83-a102-7ca37a13c7d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.033077] env[61936]: DEBUG oslo_vmware.api [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1028.033077] env[61936]: value = "task-1253392" [ 1028.033077] env[61936]: _type = "Task" [ 1028.033077] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.037285] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1028.037285] env[61936]: value = "task-1253393" [ 1028.037285] env[61936]: _type = "Task" [ 1028.037285] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.043178] env[61936]: DEBUG oslo_vmware.api [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.048314] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.093879] env[61936]: DEBUG oslo_concurrency.lockutils [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.291303] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253391, 'name': Rename_Task, 'duration_secs': 0.169384} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.291303] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1028.291746] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-403a21a2-aafa-4422-8ed1-8e446383692a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.299566] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1028.299566] env[61936]: value = "task-1253394" [ 1028.299566] env[61936]: _type = "Task" [ 1028.299566] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.309795] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.396519] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d5faad7c-9b91-45c1-a094-766d0ad5f809 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.098s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.397206] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.578s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.397464] env[61936]: INFO nova.compute.manager [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Unshelving [ 1028.485638] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1028.485638] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4866ef8f-e858-41e9-aa86-816cda118b5d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.493919] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1028.493919] env[61936]: value = "task-1253395" [ 1028.493919] env[61936]: _type = "Task" [ 1028.493919] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.504665] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.549293] env[61936]: DEBUG oslo_vmware.api [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253392, 'name': PowerOnVM_Task} progress is 93%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.553278] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253393, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44873} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.553574] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 53605e96-a4d1-4a7b-8fef-1f1878b099c3/53605e96-a4d1-4a7b-8fef-1f1878b099c3.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1028.553812] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1028.554144] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b8d63db-ac79-4f6c-9624-f369590f2b70 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.567592] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1028.567592] env[61936]: value = "task-1253396" [ 1028.567592] env[61936]: _type = "Task" [ 1028.567592] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.579887] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253396, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.609644] env[61936]: DEBUG nova.objects.instance [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'pci_requests' on Instance uuid 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.813563] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253394, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.006427] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253395, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.046908] env[61936]: DEBUG oslo_vmware.api [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253392, 'name': PowerOnVM_Task, 'duration_secs': 0.745457} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.047018] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Resumed the VM {{(pid=61936) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1029.047162] env[61936]: DEBUG nova.compute.manager [None req-86a764de-255f-4259-9baa-aa5e902c8896 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1029.047928] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd41d237-dc62-4047-ac13-e84fde3e0f9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.076470] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137087} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.078897] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.079963] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06255e0-f67c-4923-9f9c-40e314cc9e84 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.104343] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 53605e96-a4d1-4a7b-8fef-1f1878b099c3/53605e96-a4d1-4a7b-8fef-1f1878b099c3.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.106387] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1029.106614] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fe68290-cee9-4987-9e27-cbc997636a16 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.121529] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c85b4d5-b56f-4281-8720-f1b796d3b5e7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.123883] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4b909ef-090b-4413-bac4-16c84ac491ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.125920] env[61936]: DEBUG nova.objects.base [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Object Instance<0137305a-dace-4eda-9d90-7233c001176f> lazy-loaded attributes: flavor,pci_requests {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1029.126138] env[61936]: DEBUG nova.network.neutron [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.135155] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff34511d-587f-4ac9-8500-ec93d6d039b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.138233] env[61936]: DEBUG oslo_vmware.api [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1029.138233] env[61936]: value = "task-1253397" [ 1029.138233] env[61936]: _type = "Task" [ 1029.138233] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.139557] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1029.139557] env[61936]: value = "task-1253398" [ 1029.139557] env[61936]: _type = "Task" [ 1029.139557] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.174466] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a0bfce-8c61-4351-938d-4f1bed67bff1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.181061] env[61936]: DEBUG oslo_vmware.api [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253397, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.185838] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.189022] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2360bb45-bbed-4758-ac2f-a5fbf4837029 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.203379] env[61936]: DEBUG nova.compute.provider_tree [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.247813] env[61936]: DEBUG nova.policy [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.310487] env[61936]: DEBUG oslo_vmware.api [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253394, 'name': PowerOnVM_Task, 'duration_secs': 0.549714} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.310841] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1029.311057] env[61936]: INFO nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Took 10.39 seconds to spawn the instance on the hypervisor. [ 1029.311243] env[61936]: DEBUG nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1029.311987] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0e207e-4981-418a-9b3c-009cec333849 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.407193] env[61936]: DEBUG nova.compute.utils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.505290] env[61936]: DEBUG oslo_vmware.api [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253395, 'name': PowerOnVM_Task, 'duration_secs': 0.526212} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.505555] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1029.505750] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-5cc01945-0989-4f31-ba09-c727540bbc1c tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance '755c5fea-e3d2-4cb3-b717-5f862401d7e1' progress to 100 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1029.657706] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.661279] env[61936]: DEBUG oslo_vmware.api [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253397, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.706580] env[61936]: DEBUG nova.scheduler.client.report [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1029.830482] env[61936]: INFO nova.compute.manager [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Took 19.65 seconds to build instance. [ 1029.913621] env[61936]: INFO nova.virt.block_device [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Booting with volume 41f3c9d7-6e22-4c69-9d06-47b73b138c6e at /dev/sdb [ 1029.956132] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80a98f18-ff3c-4afb-a06c-a4e7e00a6f91 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.967055] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec71ba5c-dbca-4639-a621-6fddf75b86b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.998983] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aecb819f-b4b7-4e3a-8d3f-718d7b4e1be1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.008162] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baff012b-d6f4-45f6-a0c6-592d97a1836e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.042319] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f981b91-e976-4a9c-bf8e-39a4c1393f43 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.049386] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8e4e69-6c55-4db9-81a8-b43e7baac704 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.065836] env[61936]: DEBUG nova.virt.block_device [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating existing volume attachment record: ffb8c048-58e9-4369-956a-c5e63950f039 {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1030.160865] env[61936]: DEBUG oslo_vmware.api [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253397, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.164238] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.212104] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.212733] env[61936]: DEBUG nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1030.332573] env[61936]: DEBUG oslo_concurrency.lockutils [None req-55cd0c75-64a4-4d19-ac84-a6f1b899fe94 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "1c391c45-a041-422d-9a44-a29306f99a6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.158s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.657479] env[61936]: DEBUG oslo_vmware.api [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253397, 'name': PowerOnVM_Task, 'duration_secs': 1.177334} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.660698] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.660998] env[61936]: DEBUG nova.compute.manager [None req-6fd5ea79-48ad-4ad2-9f6a-4086195f9b3e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1030.661285] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253398, 'name': ReconfigVM_Task, 'duration_secs': 1.211936} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.661999] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263ec6fc-0091-4aff-b4e9-170503bd4123 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.665146] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 53605e96-a4d1-4a7b-8fef-1f1878b099c3/53605e96-a4d1-4a7b-8fef-1f1878b099c3.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.665783] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f822576-d930-4d1e-b229-1d3415bd1dab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.677235] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1030.677235] env[61936]: value = "task-1253402" [ 1030.677235] env[61936]: _type = "Task" [ 1030.677235] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.690076] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253402, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.717721] env[61936]: DEBUG nova.compute.utils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.719029] env[61936]: DEBUG nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1030.719189] env[61936]: DEBUG nova.network.neutron [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1030.771564] env[61936]: DEBUG nova.policy [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b05bdcace008487dba580405eb672ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90c29449db6c4beaa7f190d9225cb08a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1030.884027] env[61936]: DEBUG nova.network.neutron [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Successfully updated port: fa637c42-69ec-480c-bc7b-08a8a723660b {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.040935] env[61936]: DEBUG nova.network.neutron [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Successfully created port: b56dc652-5710-432a-a37f-d8598b217179 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1031.194037] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253402, 'name': Rename_Task, 'duration_secs': 0.30164} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.194037] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.194852] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd104a47-d1a7-40b4-bfa2-695bf335a1be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.202635] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1031.202635] env[61936]: value = "task-1253403" [ 1031.202635] env[61936]: _type = "Task" [ 1031.202635] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.212667] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.222867] env[61936]: DEBUG nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1031.387313] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.387747] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.387747] env[61936]: DEBUG nova.network.neutron [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.422815] env[61936]: DEBUG nova.compute.manager [req-7827e919-90bc-46f0-8164-a446f50f58e5 req-44d08705-e2b5-4210-aba4-d004eb5e84a4 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-vif-plugged-fa637c42-69ec-480c-bc7b-08a8a723660b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1031.423423] env[61936]: DEBUG oslo_concurrency.lockutils [req-7827e919-90bc-46f0-8164-a446f50f58e5 req-44d08705-e2b5-4210-aba4-d004eb5e84a4 service nova] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.423768] env[61936]: DEBUG oslo_concurrency.lockutils [req-7827e919-90bc-46f0-8164-a446f50f58e5 req-44d08705-e2b5-4210-aba4-d004eb5e84a4 service nova] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.423972] env[61936]: DEBUG oslo_concurrency.lockutils [req-7827e919-90bc-46f0-8164-a446f50f58e5 req-44d08705-e2b5-4210-aba4-d004eb5e84a4 service nova] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.424321] env[61936]: DEBUG nova.compute.manager [req-7827e919-90bc-46f0-8164-a446f50f58e5 req-44d08705-e2b5-4210-aba4-d004eb5e84a4 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] No waiting events found dispatching network-vif-plugged-fa637c42-69ec-480c-bc7b-08a8a723660b {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1031.424718] env[61936]: WARNING nova.compute.manager [req-7827e919-90bc-46f0-8164-a446f50f58e5 req-44d08705-e2b5-4210-aba4-d004eb5e84a4 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received unexpected event network-vif-plugged-fa637c42-69ec-480c-bc7b-08a8a723660b for instance with vm_state active and task_state None. [ 1031.714967] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253403, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.797757] env[61936]: DEBUG nova.compute.manager [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1031.798747] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26afaef-7110-451b-94f5-2f41a298108f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.862071] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "484e5d41-29f4-4845-9633-157c03766978" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.862407] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.926061] env[61936]: WARNING nova.network.neutron [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] 03166185-d5e0-477b-892e-20bea284c8db already exists in list: networks containing: ['03166185-d5e0-477b-892e-20bea284c8db']. ignoring it [ 1031.926258] env[61936]: WARNING nova.network.neutron [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] 03166185-d5e0-477b-892e-20bea284c8db already exists in list: networks containing: ['03166185-d5e0-477b-892e-20bea284c8db']. ignoring it [ 1032.161061] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.161561] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.161561] env[61936]: DEBUG nova.compute.manager [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Going to confirm migration 2 {{(pid=61936) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5174}} [ 1032.213712] env[61936]: DEBUG oslo_vmware.api [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253403, 'name': PowerOnVM_Task, 'duration_secs': 0.565243} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.214138] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1032.214240] env[61936]: INFO nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Took 9.29 seconds to spawn the instance on the hypervisor. [ 1032.214389] env[61936]: DEBUG nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1032.215215] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2abfe0-a558-4bd5-a1a9-978d51407fdd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.231997] env[61936]: DEBUG nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1032.254525] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1032.254786] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.254955] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.255164] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.255315] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.255463] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1032.255697] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1032.255903] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1032.256105] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1032.256282] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1032.256472] env[61936]: DEBUG nova.virt.hardware [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1032.257677] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70687113-8d5c-4ab3-a6eb-08a2fcf47aa5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.267640] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0480fa96-a1b0-4c4f-80b8-5c3789721ae4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.315088] env[61936]: INFO nova.compute.manager [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] instance snapshotting [ 1032.315088] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d3e9d7-99c6-45d7-8a74-b7825c0cc870 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.338880] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51e789c-e5b0-4145-8ddc-64dfcd8431fd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.366162] env[61936]: DEBUG nova.compute.utils [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1032.383772] env[61936]: DEBUG nova.network.neutron [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "address": "fa:16:3e:fe:29:21", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ccd46c0-94", "ovs_interfaceid": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fa637c42-69ec-480c-bc7b-08a8a723660b", "address": "fa:16:3e:d0:9d:1d", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa637c42-69", "ovs_interfaceid": "fa637c42-69ec-480c-bc7b-08a8a723660b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.571632] env[61936]: DEBUG nova.network.neutron [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Successfully updated port: b56dc652-5710-432a-a37f-d8598b217179 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.733843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.734094] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.734305] env[61936]: DEBUG nova.network.neutron [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.734490] env[61936]: DEBUG nova.objects.instance [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lazy-loading 'info_cache' on Instance uuid 755c5fea-e3d2-4cb3-b717-5f862401d7e1 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.737718] env[61936]: INFO nova.compute.manager [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Took 18.52 seconds to build instance. [ 1032.855032] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Creating Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1032.855032] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-af43ac81-b240-4a4a-9371-d00ef3747b82 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.863016] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1032.863016] env[61936]: value = "task-1253405" [ 1032.863016] env[61936]: _type = "Task" [ 1032.863016] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.871919] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.872566] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253405, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.886779] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.887497] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.887687] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.888539] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18497bf7-d5b4-4875-8b0a-7c55fc3bed98 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.906377] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1032.906630] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.906793] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.907332] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.907332] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.907332] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1032.907504] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1032.907625] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1032.907790] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1032.907955] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1032.908150] env[61936]: DEBUG nova.virt.hardware [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1032.914336] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfiguring VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1032.914656] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f16cc01d-bbc6-457f-97c0-0bb61a91f397 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.933229] env[61936]: DEBUG oslo_vmware.api [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1032.933229] env[61936]: value = "task-1253406" [ 1032.933229] env[61936]: _type = "Task" [ 1032.933229] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.941834] env[61936]: DEBUG oslo_vmware.api [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253406, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.078351] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-b4fb4cfc-10a5-4b24-9f79-1e6832659f89" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.078501] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-b4fb4cfc-10a5-4b24-9f79-1e6832659f89" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.078709] env[61936]: DEBUG nova.network.neutron [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.240957] env[61936]: DEBUG oslo_concurrency.lockutils [None req-38bbc7f9-10ae-4220-9447-e3345e9af32e tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.027s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.375727] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253405, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.443842] env[61936]: DEBUG oslo_vmware.api [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253406, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.623712] env[61936]: DEBUG nova.network.neutron [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1033.787723] env[61936]: DEBUG nova.network.neutron [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Updating instance_info_cache with network_info: [{"id": "b56dc652-5710-432a-a37f-d8598b217179", "address": "fa:16:3e:89:02:1b", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb56dc652-57", "ovs_interfaceid": "b56dc652-5710-432a-a37f-d8598b217179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.876741] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253405, 'name': CreateSnapshot_Task, 'duration_secs': 0.808374} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.877068] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Created Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1033.877811] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b290ebd4-16f3-4b61-992c-f1d6d6c8a554 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.934799] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "484e5d41-29f4-4845-9633-157c03766978" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.935357] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.935457] env[61936]: INFO nova.compute.manager [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Attaching volume 0daa7ba3-e820-4de6-a26e-a0fecebe9746 to /dev/sdb [ 1033.958167] env[61936]: DEBUG oslo_vmware.api [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253406, 'name': ReconfigVM_Task, 'duration_secs': 0.878675} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.958496] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.958716] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfigured VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1033.986380] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01269fa3-a62f-4d7e-a5db-324cf855af06 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.996574] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20e6fbb-b51b-4e7b-b467-6d5efb183be9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.013239] env[61936]: DEBUG nova.virt.block_device [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating existing volume attachment record: 39302dc8-b4df-43d7-b80a-48b2b06944d8 {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1034.099525] env[61936]: DEBUG nova.network.neutron [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance_info_cache with network_info: [{"id": "7775677e-cf35-458b-8c94-da7dd8dc652a", "address": "fa:16:3e:f7:93:44", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7775677e-cf", "ovs_interfaceid": "7775677e-cf35-458b-8c94-da7dd8dc652a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.202279] env[61936]: DEBUG nova.compute.manager [req-53ee3afb-8c30-4678-b4bb-a5234b18ce6a req-c78dbfd2-8083-4290-b26b-8425b3366218 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Received event network-vif-plugged-b56dc652-5710-432a-a37f-d8598b217179 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1034.202279] env[61936]: DEBUG oslo_concurrency.lockutils [req-53ee3afb-8c30-4678-b4bb-a5234b18ce6a req-c78dbfd2-8083-4290-b26b-8425b3366218 service nova] Acquiring lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.202496] env[61936]: DEBUG oslo_concurrency.lockutils [req-53ee3afb-8c30-4678-b4bb-a5234b18ce6a req-c78dbfd2-8083-4290-b26b-8425b3366218 service nova] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.202624] env[61936]: DEBUG oslo_concurrency.lockutils [req-53ee3afb-8c30-4678-b4bb-a5234b18ce6a req-c78dbfd2-8083-4290-b26b-8425b3366218 service nova] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.202797] env[61936]: DEBUG nova.compute.manager [req-53ee3afb-8c30-4678-b4bb-a5234b18ce6a req-c78dbfd2-8083-4290-b26b-8425b3366218 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] No waiting events found dispatching network-vif-plugged-b56dc652-5710-432a-a37f-d8598b217179 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1034.202965] env[61936]: WARNING nova.compute.manager [req-53ee3afb-8c30-4678-b4bb-a5234b18ce6a req-c78dbfd2-8083-4290-b26b-8425b3366218 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Received unexpected event network-vif-plugged-b56dc652-5710-432a-a37f-d8598b217179 for instance with vm_state building and task_state spawning. [ 1034.220901] env[61936]: DEBUG nova.compute.manager [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-changed-fa637c42-69ec-480c-bc7b-08a8a723660b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1034.220901] env[61936]: DEBUG nova.compute.manager [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing instance network info cache due to event network-changed-fa637c42-69ec-480c-bc7b-08a8a723660b. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1034.220901] env[61936]: DEBUG oslo_concurrency.lockutils [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.220901] env[61936]: DEBUG oslo_concurrency.lockutils [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.221234] env[61936]: DEBUG nova.network.neutron [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Refreshing network info cache for port fa637c42-69ec-480c-bc7b-08a8a723660b {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1034.290216] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-b4fb4cfc-10a5-4b24-9f79-1e6832659f89" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.290607] env[61936]: DEBUG nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Instance network_info: |[{"id": "b56dc652-5710-432a-a37f-d8598b217179", "address": "fa:16:3e:89:02:1b", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb56dc652-57", "ovs_interfaceid": "b56dc652-5710-432a-a37f-d8598b217179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1034.291094] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:02:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b56dc652-5710-432a-a37f-d8598b217179', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.298929] env[61936]: DEBUG oslo.service.loopingcall [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.300023] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1034.300278] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a59bf469-9ee4-4a4e-9b30-cd68c4b90093 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.320593] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.320593] env[61936]: value = "task-1253410" [ 1034.320593] env[61936]: _type = "Task" [ 1034.320593] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.331377] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253410, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.396821] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Creating linked-clone VM from snapshot {{(pid=61936) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1034.397165] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b10e150e-f3f6-4c07-a856-730bbea3e61a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.407081] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1034.407081] env[61936]: value = "task-1253411" [ 1034.407081] env[61936]: _type = "Task" [ 1034.407081] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.416268] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253411, 'name': CloneVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.464146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cf005c37-1226-405f-88e0-bcbce8906c23 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-fa637c42-69ec-480c-bc7b-08a8a723660b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.492s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.602948] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-755c5fea-e3d2-4cb3-b717-5f862401d7e1" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.603302] env[61936]: DEBUG nova.objects.instance [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lazy-loading 'migration_context' on Instance uuid 755c5fea-e3d2-4cb3-b717-5f862401d7e1 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.833774] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253410, 'name': CreateVM_Task, 'duration_secs': 0.483796} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.834053] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.838880] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.839095] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.839467] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.842272] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a520a54-d43f-44d8-91e1-abb8cb3879fa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.848474] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1034.848474] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232d581-a749-e25e-7329-c73ae4a312f0" [ 1034.848474] env[61936]: _type = "Task" [ 1034.848474] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.857663] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232d581-a749-e25e-7329-c73ae4a312f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.918596] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253411, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.003513] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.003513] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.004301] env[61936]: DEBUG nova.compute.manager [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1035.005441] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f031051-d1cd-4b59-aefa-47ec09171c36 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.014410] env[61936]: DEBUG nova.compute.manager [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61936) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1035.014831] env[61936]: DEBUG nova.objects.instance [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'flavor' on Instance uuid 53605e96-a4d1-4a7b-8fef-1f1878b099c3 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.106554] env[61936]: DEBUG nova.objects.base [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Object Instance<755c5fea-e3d2-4cb3-b717-5f862401d7e1> lazy-loaded attributes: info_cache,migration_context {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1035.107939] env[61936]: DEBUG nova.network.neutron [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updated VIF entry in instance network info cache for port fa637c42-69ec-480c-bc7b-08a8a723660b. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1035.108568] env[61936]: DEBUG nova.network.neutron [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "address": "fa:16:3e:fe:29:21", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ccd46c0-94", "ovs_interfaceid": "3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fa637c42-69ec-480c-bc7b-08a8a723660b", "address": "fa:16:3e:d0:9d:1d", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa637c42-69", "ovs_interfaceid": "fa637c42-69ec-480c-bc7b-08a8a723660b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.111040] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b12a3ee-290c-4d04-8b19-aa1d30ed3b6e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.137375] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65894331-872f-4571-817c-641136bbaac2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.145011] env[61936]: DEBUG oslo_vmware.api [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1035.145011] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529e226d-75a3-827e-e6eb-cafb150b1250" [ 1035.145011] env[61936]: _type = "Task" [ 1035.145011] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.155986] env[61936]: DEBUG oslo_vmware.api [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529e226d-75a3-827e-e6eb-cafb150b1250, 'name': SearchDatastore_Task, 'duration_secs': 0.009031} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.159052] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.159052] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.360049] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232d581-a749-e25e-7329-c73ae4a312f0, 'name': SearchDatastore_Task, 'duration_secs': 0.011667} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.360749] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.360749] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.361071] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.361245] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.361503] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.361814] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12d7dc69-fc33-4180-9c40-9e062558d1dd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.371776] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.371996] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1035.372915] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28ed43d6-187d-4860-ba5a-56fd1051d45a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.378519] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1035.378519] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528ef0e7-3375-c07c-3470-9faf9a9c8b6e" [ 1035.378519] env[61936]: _type = "Task" [ 1035.378519] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.387827] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528ef0e7-3375-c07c-3470-9faf9a9c8b6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.419515] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253411, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.616326] env[61936]: DEBUG oslo_concurrency.lockutils [req-175c9afa-3d25-47d7-a519-3e68908179f3 req-91db3c28-f5fb-432a-bc62-9d9d6a313e47 service nova] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.684729] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.846153] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905fa6e8-a095-4f44-8a9d-917959b89af3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.854694] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e584ae-ef65-4ad9-9a14-f7016268e77c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.889064] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f590f9-16e4-4e04-af6e-21000f840a70 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.900492] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c400cc6f-e600-4f9d-ad1d-0f95f25d67c6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.904205] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528ef0e7-3375-c07c-3470-9faf9a9c8b6e, 'name': SearchDatastore_Task, 'duration_secs': 0.012086} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.905322] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eee355c8-feff-482a-93a7-e5dca89de7ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.917095] env[61936]: DEBUG nova.compute.provider_tree [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.924647] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1035.924647] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232b6b2-a21c-a149-ce6f-fadb3c73ce83" [ 1035.924647] env[61936]: _type = "Task" [ 1035.924647] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.928274] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253411, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.938242] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232b6b2-a21c-a149-ce6f-fadb3c73ce83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.022669] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.023088] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe42eead-b84f-4213-9d45-5c6fd4ab1963 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.031882] env[61936]: DEBUG oslo_vmware.api [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1036.031882] env[61936]: value = "task-1253412" [ 1036.031882] env[61936]: _type = "Task" [ 1036.031882] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.041333] env[61936]: DEBUG oslo_vmware.api [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253412, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.074837] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-0137305a-dace-4eda-9d90-7233c001176f-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.075134] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.283011] env[61936]: DEBUG nova.compute.manager [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Received event network-changed-b56dc652-5710-432a-a37f-d8598b217179 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1036.283298] env[61936]: DEBUG nova.compute.manager [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Refreshing instance network info cache due to event network-changed-b56dc652-5710-432a-a37f-d8598b217179. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1036.283584] env[61936]: DEBUG oslo_concurrency.lockutils [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] Acquiring lock "refresh_cache-b4fb4cfc-10a5-4b24-9f79-1e6832659f89" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.283736] env[61936]: DEBUG oslo_concurrency.lockutils [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] Acquired lock "refresh_cache-b4fb4cfc-10a5-4b24-9f79-1e6832659f89" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.283903] env[61936]: DEBUG nova.network.neutron [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Refreshing network info cache for port b56dc652-5710-432a-a37f-d8598b217179 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1036.423828] env[61936]: DEBUG nova.scheduler.client.report [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1036.432372] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253411, 'name': CloneVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.443117] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5232b6b2-a21c-a149-ce6f-fadb3c73ce83, 'name': SearchDatastore_Task, 'duration_secs': 0.01293} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.444225] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.444542] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] b4fb4cfc-10a5-4b24-9f79-1e6832659f89/b4fb4cfc-10a5-4b24-9f79-1e6832659f89.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1036.444779] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e308f8be-7c33-428f-85c4-f04f6239cdae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.452873] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1036.452873] env[61936]: value = "task-1253414" [ 1036.452873] env[61936]: _type = "Task" [ 1036.452873] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.462372] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253414, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.542455] env[61936]: DEBUG oslo_vmware.api [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253412, 'name': PowerOffVM_Task, 'duration_secs': 0.287238} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.542816] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.543049] env[61936]: DEBUG nova.compute.manager [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1036.543971] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d65a02-c45e-4ffb-973d-cd3b5c1a9336 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.579339] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.579682] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.581757] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae7ba99-5ff5-4d59-b4bb-b85a6ac68e9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.612790] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-007039a3-b0cf-40d8-8938-b095f58e0dfc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.658513] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfiguring VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1036.659507] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1027d143-7901-4fa8-aa6e-f06477edbf14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.686898] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1036.686898] env[61936]: value = "task-1253415" [ 1036.686898] env[61936]: _type = "Task" [ 1036.686898] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.698052] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.933492] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253411, 'name': CloneVM_Task, 'duration_secs': 2.03199} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.936920] env[61936]: INFO nova.virt.vmwareapi.vmops [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Created linked-clone VM from snapshot [ 1036.937857] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c311f66-6c92-46a2-9257-7bc8962c3d08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.948595] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Uploading image 4f39bc9e-7ede-4516-87cd-fe1cc675b904 {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1036.964842] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253414, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501392} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.965769] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Destroying the VM {{(pid=61936) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1036.966079] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] b4fb4cfc-10a5-4b24-9f79-1e6832659f89/b4fb4cfc-10a5-4b24-9f79-1e6832659f89.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1036.966290] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.966510] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7d118b35-c5e8-46db-8c17-075d6bf0589b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.968050] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10b30be7-9553-4c83-a8ac-7f1dff3dc0a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.975583] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1036.975583] env[61936]: value = "task-1253417" [ 1036.975583] env[61936]: _type = "Task" [ 1036.975583] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.978934] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1036.978934] env[61936]: value = "task-1253416" [ 1036.978934] env[61936]: _type = "Task" [ 1036.978934] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.993912] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253416, 'name': Destroy_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.998924] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253417, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.018772] env[61936]: DEBUG nova.network.neutron [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Updated VIF entry in instance network info cache for port b56dc652-5710-432a-a37f-d8598b217179. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.019191] env[61936]: DEBUG nova.network.neutron [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Updating instance_info_cache with network_info: [{"id": "b56dc652-5710-432a-a37f-d8598b217179", "address": "fa:16:3e:89:02:1b", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb56dc652-57", "ovs_interfaceid": "b56dc652-5710-432a-a37f-d8598b217179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.056887] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1e0d0188-102a-4c64-93f9-c5521d52506f tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.197055] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.435990] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.277s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.439541] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.755s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.439951] env[61936]: DEBUG nova.objects.instance [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'pci_requests' on Instance uuid fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.492749] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253417, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070381} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.493516] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.495008] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b429a7-9cc3-4c0d-ad93-6e2d8117ce45 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.501268] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253416, 'name': Destroy_Task, 'duration_secs': 0.334344} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.501885] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Destroyed the VM [ 1037.502213] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Deleting Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1037.502467] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ce1576b8-d3ad-471b-b952-2e77b2346726 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.522707] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] b4fb4cfc-10a5-4b24-9f79-1e6832659f89/b4fb4cfc-10a5-4b24-9f79-1e6832659f89.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.523818] env[61936]: DEBUG oslo_concurrency.lockutils [req-8a046e48-f9a9-489c-b26a-e6c9be5f1996 req-3e45ba69-6175-4997-9123-1a22f84bba15 service nova] Releasing lock "refresh_cache-b4fb4cfc-10a5-4b24-9f79-1e6832659f89" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.524511] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39530239-8e7e-4be8-9804-70ae07f9c9ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.545927] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1037.545927] env[61936]: value = "task-1253418" [ 1037.545927] env[61936]: _type = "Task" [ 1037.545927] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.556619] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253418, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.558303] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1037.558303] env[61936]: value = "task-1253419" [ 1037.558303] env[61936]: _type = "Task" [ 1037.558303] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.567657] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253419, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.700349] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.792955] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.793271] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.793496] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.793711] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.793864] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.796459] env[61936]: INFO nova.compute.manager [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Terminating instance [ 1037.946047] env[61936]: DEBUG nova.objects.instance [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'numa_topology' on Instance uuid fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.001011] env[61936]: INFO nova.scheduler.client.report [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted allocation for migration 2d0e61d3-548f-4a27-a434-95d93432350b [ 1038.058580] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253418, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.067566] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253419, 'name': ReconfigVM_Task, 'duration_secs': 0.510533} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.067942] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Reconfigured VM instance instance-00000063 to attach disk [datastore1] b4fb4cfc-10a5-4b24-9f79-1e6832659f89/b4fb4cfc-10a5-4b24-9f79-1e6832659f89.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.068650] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e5d168c-556d-49a0-8c53-f117b63e3a4a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.078038] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1038.078038] env[61936]: value = "task-1253420" [ 1038.078038] env[61936]: _type = "Task" [ 1038.078038] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.087870] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253420, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.202599] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.301767] env[61936]: DEBUG nova.compute.manager [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1038.302031] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.302981] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c3ef74-d746-4047-a969-166fd14fa49e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.311893] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.312209] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0bf0ca7-96c0-46b8-859b-9453d789741a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.375188] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.375422] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.375607] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleting the datastore file [datastore1] 53605e96-a4d1-4a7b-8fef-1f1878b099c3 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.375884] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1be9aa52-e977-460f-b5bb-2bc5ec2e71c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.382925] env[61936]: DEBUG oslo_vmware.api [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1038.382925] env[61936]: value = "task-1253422" [ 1038.382925] env[61936]: _type = "Task" [ 1038.382925] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.391754] env[61936]: DEBUG oslo_vmware.api [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.450693] env[61936]: INFO nova.compute.claims [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.508018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7eb0ea2-3b9d-4e5b-9d88-3a84ebb5e2ea tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.346s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.558462] env[61936]: DEBUG oslo_vmware.api [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253418, 'name': RemoveSnapshot_Task, 'duration_secs': 0.78205} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.558675] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Deleted Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1038.563804] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1038.564191] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270028', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'name': 'volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '484e5d41-29f4-4845-9633-157c03766978', 'attached_at': '', 'detached_at': '', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'serial': '0daa7ba3-e820-4de6-a26e-a0fecebe9746'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1038.565103] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8b2700-3519-45d6-aaa8-531600046bf6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.588856] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d88ab85-4f09-4b5c-949a-25ef0218acec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.614862] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253420, 'name': Rename_Task, 'duration_secs': 0.169691} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.623602] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746/volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.623972] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1038.624810] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e61ef8e3-f5e5-45c3-91e2-225f4a49e88f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.638347] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecf652fb-7d5e-47eb-8a3e-2fecbcbd76f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.647619] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1038.647619] env[61936]: value = "task-1253423" [ 1038.647619] env[61936]: _type = "Task" [ 1038.647619] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.649146] env[61936]: DEBUG oslo_vmware.api [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1038.649146] env[61936]: value = "task-1253424" [ 1038.649146] env[61936]: _type = "Task" [ 1038.649146] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.663604] env[61936]: DEBUG oslo_vmware.api [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253424, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.663853] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253423, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.706073] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.894406] env[61936]: DEBUG oslo_vmware.api [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179849} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.894697] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.894873] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.895064] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.895245] env[61936]: INFO nova.compute.manager [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1038.895488] env[61936]: DEBUG oslo.service.loopingcall [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.895746] env[61936]: DEBUG nova.compute.manager [-] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1038.895870] env[61936]: DEBUG nova.network.neutron [-] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.038338] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.038664] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.038884] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.039215] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.039586] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.041867] env[61936]: INFO nova.compute.manager [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Terminating instance [ 1039.064564] env[61936]: WARNING nova.compute.manager [None req-125d716e-7f84-4565-a152-431c33a5d7e4 tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Image not found during snapshot: nova.exception.ImageNotFound: Image 4f39bc9e-7ede-4516-87cd-fe1cc675b904 could not be found. [ 1039.162702] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253423, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.165875] env[61936]: DEBUG oslo_vmware.api [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253424, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.200578] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.545539] env[61936]: DEBUG nova.compute.manager [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1039.545775] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1039.547469] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50241ac-e34a-46ca-9aa8-91f100ce9af0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.558316] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1039.558600] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-687e7c3b-ad9c-41da-89d5-a32bfd6749c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.562009] env[61936]: DEBUG nova.compute.manager [req-e237f430-0de1-444c-8570-5aeee74928c0 req-f335ab57-60e7-46ff-8e34-6ee7c8c446b6 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Received event network-vif-deleted-69b0f1f9-2468-4011-bab2-0006651f7bba {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1039.562221] env[61936]: INFO nova.compute.manager [req-e237f430-0de1-444c-8570-5aeee74928c0 req-f335ab57-60e7-46ff-8e34-6ee7c8c446b6 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Neutron deleted interface 69b0f1f9-2468-4011-bab2-0006651f7bba; detaching it from the instance and deleting it from the info cache [ 1039.563014] env[61936]: DEBUG nova.network.neutron [req-e237f430-0de1-444c-8570-5aeee74928c0 req-f335ab57-60e7-46ff-8e34-6ee7c8c446b6 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.571253] env[61936]: DEBUG oslo_vmware.api [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1039.571253] env[61936]: value = "task-1253425" [ 1039.571253] env[61936]: _type = "Task" [ 1039.571253] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.581534] env[61936]: DEBUG oslo_vmware.api [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.666411] env[61936]: DEBUG oslo_vmware.api [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253423, 'name': PowerOnVM_Task, 'duration_secs': 0.843613} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.666411] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1039.666411] env[61936]: INFO nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Took 7.43 seconds to spawn the instance on the hypervisor. [ 1039.666411] env[61936]: DEBUG nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1039.666411] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66f869c-55ea-48b5-bad1-56f34a39396f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.671476] env[61936]: DEBUG oslo_vmware.api [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253424, 'name': ReconfigVM_Task, 'duration_secs': 0.867589} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.674175] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfigured VM instance instance-0000005b to attach disk [datastore1] volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746/volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.681330] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79819d58-13b3-4b56-bf76-8c40af9c51f1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.693981] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed0f325-b21a-4c3c-9150-1b6cb53ea4f0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.701985] env[61936]: DEBUG oslo_vmware.api [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1039.701985] env[61936]: value = "task-1253426" [ 1039.701985] env[61936]: _type = "Task" [ 1039.701985] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.717476] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc20b61-8270-46a8-9bd3-40372172b679 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.721401] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.727928] env[61936]: DEBUG oslo_vmware.api [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253426, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.753836] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f813941-d5a4-40fc-a1d7-a75d32d5250b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.762648] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1d6268-dd87-4157-bb71-22e7278aab3a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.776908] env[61936]: DEBUG nova.compute.provider_tree [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.979967] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "1c391c45-a041-422d-9a44-a29306f99a6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.979967] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "1c391c45-a041-422d-9a44-a29306f99a6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.980178] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "1c391c45-a041-422d-9a44-a29306f99a6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.980371] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "1c391c45-a041-422d-9a44-a29306f99a6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.980543] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "1c391c45-a041-422d-9a44-a29306f99a6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.983216] env[61936]: INFO nova.compute.manager [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Terminating instance [ 1040.015158] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.015436] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.015647] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.015850] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.016075] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.018539] env[61936]: INFO nova.compute.manager [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Terminating instance [ 1040.035891] env[61936]: DEBUG nova.network.neutron [-] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.067027] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86b61920-fdff-44d6-8b2d-8fb27cc9cc8d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.079061] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a01062-c893-4ed6-b805-698cbc2ec36b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.100044] env[61936]: DEBUG oslo_vmware.api [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253425, 'name': PowerOffVM_Task, 'duration_secs': 0.215647} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.100044] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1040.100044] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1040.100214] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-072a9a7a-e6be-4065-aaf8-7e632f6913fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.121475] env[61936]: DEBUG nova.compute.manager [req-e237f430-0de1-444c-8570-5aeee74928c0 req-f335ab57-60e7-46ff-8e34-6ee7c8c446b6 service nova] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Detach interface failed, port_id=69b0f1f9-2468-4011-bab2-0006651f7bba, reason: Instance 53605e96-a4d1-4a7b-8fef-1f1878b099c3 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1040.205618] env[61936]: INFO nova.compute.manager [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Took 14.23 seconds to build instance. [ 1040.208247] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1040.208647] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1040.208714] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleting the datastore file [datastore2] 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.209051] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3d84e9e-439f-4512-bab1-675ab19b2b43 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.217684] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.223379] env[61936]: DEBUG oslo_vmware.api [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253426, 'name': ReconfigVM_Task, 'duration_secs': 0.158052} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.224629] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270028', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'name': 'volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '484e5d41-29f4-4845-9633-157c03766978', 'attached_at': '', 'detached_at': '', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'serial': '0daa7ba3-e820-4de6-a26e-a0fecebe9746'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1040.226027] env[61936]: DEBUG oslo_vmware.api [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for the task: (returnval){ [ 1040.226027] env[61936]: value = "task-1253428" [ 1040.226027] env[61936]: _type = "Task" [ 1040.226027] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.233973] env[61936]: DEBUG oslo_vmware.api [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.280397] env[61936]: DEBUG nova.scheduler.client.report [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1040.486833] env[61936]: DEBUG nova.compute.manager [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1040.487086] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.488045] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9a24c9-2cef-4fd5-a5cd-8174e559b916 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.496039] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.496288] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75411382-e06b-490f-9d22-dcf1b223bd73 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.502243] env[61936]: DEBUG oslo_vmware.api [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1040.502243] env[61936]: value = "task-1253429" [ 1040.502243] env[61936]: _type = "Task" [ 1040.502243] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.509803] env[61936]: DEBUG oslo_vmware.api [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.521934] env[61936]: DEBUG nova.compute.manager [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1040.522345] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.524539] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce2c5fd-2a7e-46a5-9e13-bca376ecac61 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.530365] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.530605] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8224dd28-a42e-44ac-b701-d984cb27e872 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.537075] env[61936]: DEBUG oslo_vmware.api [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1040.537075] env[61936]: value = "task-1253430" [ 1040.537075] env[61936]: _type = "Task" [ 1040.537075] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.537536] env[61936]: INFO nova.compute.manager [-] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Took 1.64 seconds to deallocate network for instance. [ 1040.549989] env[61936]: DEBUG oslo_vmware.api [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.600985] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "160d7915-3795-460d-961f-7b43c8ed0168" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.601258] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "160d7915-3795-460d-961f-7b43c8ed0168" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.705231] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.708007] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a49009ef-68a5-407b-b11a-5d9ce8bc57fa tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.743s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.743929] env[61936]: DEBUG oslo_vmware.api [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Task: {'id': task-1253428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.39976} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.744377] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1040.744712] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1040.745060] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1040.745394] env[61936]: INFO nova.compute.manager [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1040.745776] env[61936]: DEBUG oslo.service.loopingcall [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.746081] env[61936]: DEBUG nova.compute.manager [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1040.746228] env[61936]: DEBUG nova.network.neutron [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1040.785015] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.345s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.813371] env[61936]: INFO nova.network.neutron [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating port d169ca50-5a14-4a65-9da0-ea79c80f75c8 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1040.884107] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.884405] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.884597] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.885930] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.885930] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.889023] env[61936]: INFO nova.compute.manager [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Terminating instance [ 1041.018214] env[61936]: DEBUG oslo_vmware.api [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253429, 'name': PowerOffVM_Task, 'duration_secs': 0.191611} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.018538] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.018712] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.018982] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19b483a2-56a0-4c7f-9ca8-6627ba323bc2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.049095] env[61936]: DEBUG oslo_vmware.api [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253430, 'name': PowerOffVM_Task, 'duration_secs': 0.235369} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.049168] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.049310] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.049560] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf855c0a-9ed1-4cbc-9041-3b2494e5a1a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.051994] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.052379] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.052610] env[61936]: DEBUG nova.objects.instance [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'resources' on Instance uuid 53605e96-a4d1-4a7b-8fef-1f1878b099c3 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.084343] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.084614] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.084773] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleting the datastore file [datastore1] 1c391c45-a041-422d-9a44-a29306f99a6d {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.085907] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ea91470-04d1-460f-a9c3-527f163238a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.093740] env[61936]: DEBUG oslo_vmware.api [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for the task: (returnval){ [ 1041.093740] env[61936]: value = "task-1253433" [ 1041.093740] env[61936]: _type = "Task" [ 1041.093740] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.106566] env[61936]: DEBUG nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1041.109992] env[61936]: DEBUG oslo_vmware.api [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253433, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.118072] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.118328] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.118513] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleting the datastore file [datastore1] 755c5fea-e3d2-4cb3-b717-5f862401d7e1 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.118808] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-caea4eca-d315-4722-ba28-ad88cc950890 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.125102] env[61936]: DEBUG oslo_vmware.api [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1041.125102] env[61936]: value = "task-1253434" [ 1041.125102] env[61936]: _type = "Task" [ 1041.125102] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.135865] env[61936]: DEBUG oslo_vmware.api [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253434, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.204689] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.266419] env[61936]: DEBUG nova.objects.instance [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lazy-loading 'flavor' on Instance uuid 484e5d41-29f4-4845-9633-157c03766978 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.394675] env[61936]: DEBUG nova.compute.manager [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1041.394942] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1041.395862] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c2e12a-bf0f-4785-9524-14fc6b53fb50 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.404170] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.404428] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a0bd0f1-3abd-43a9-9e84-8bdeb08e89c4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.411053] env[61936]: DEBUG oslo_vmware.api [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1041.411053] env[61936]: value = "task-1253435" [ 1041.411053] env[61936]: _type = "Task" [ 1041.411053] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.420306] env[61936]: DEBUG oslo_vmware.api [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.534489] env[61936]: INFO nova.compute.manager [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Rescuing [ 1041.534885] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.535079] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.535482] env[61936]: DEBUG nova.network.neutron [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.600132] env[61936]: DEBUG nova.compute.manager [req-68d3a539-cc3a-4eba-8654-04212076766c req-98de6e87-6602-43f9-af85-3266634b9516 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Received event network-vif-deleted-1cdd9a15-01e7-40cf-80fc-24bea0b642d3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1041.600420] env[61936]: INFO nova.compute.manager [req-68d3a539-cc3a-4eba-8654-04212076766c req-98de6e87-6602-43f9-af85-3266634b9516 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Neutron deleted interface 1cdd9a15-01e7-40cf-80fc-24bea0b642d3; detaching it from the instance and deleting it from the info cache [ 1041.600634] env[61936]: DEBUG nova.network.neutron [req-68d3a539-cc3a-4eba-8654-04212076766c req-98de6e87-6602-43f9-af85-3266634b9516 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.608510] env[61936]: DEBUG oslo_vmware.api [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Task: {'id': task-1253433, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145697} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.611606] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.611857] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.612244] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.612244] env[61936]: INFO nova.compute.manager [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1041.612512] env[61936]: DEBUG oslo.service.loopingcall [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.616103] env[61936]: DEBUG nova.compute.manager [-] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1041.616390] env[61936]: DEBUG nova.network.neutron [-] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.618119] env[61936]: DEBUG nova.network.neutron [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.636260] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.639966] env[61936]: DEBUG oslo_vmware.api [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253434, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166376} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.642595] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.642818] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.643050] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.643457] env[61936]: INFO nova.compute.manager [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1041.643530] env[61936]: DEBUG oslo.service.loopingcall [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.644481] env[61936]: DEBUG nova.compute.manager [-] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1041.644552] env[61936]: DEBUG nova.network.neutron [-] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.705531] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.771680] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c5193a-f7f2-436f-95f6-628d2d7bcde4 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.836s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.774007] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfa745e-96d0-481c-9773-29645231bd0d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.781899] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936f2db3-5bd6-419e-92ba-3a47f5432e6f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.817057] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b02d04-502e-419c-a472-7154d50d21a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.826044] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f28088-fb0d-4a59-9bb2-9ec2674ed0aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.844600] env[61936]: DEBUG nova.compute.provider_tree [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.921340] env[61936]: DEBUG oslo_vmware.api [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253435, 'name': PowerOffVM_Task, 'duration_secs': 0.175688} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.921635] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.921807] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.922076] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-700adc84-105f-40da-a3ce-0c89c095852e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.991271] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.991644] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.992692] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleting the datastore file [datastore1] b4fb4cfc-10a5-4b24-9f79-1e6832659f89 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.992982] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a5f503c-de81-4cdd-a25b-1c91f81501f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.004018] env[61936]: DEBUG oslo_vmware.api [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1042.004018] env[61936]: value = "task-1253437" [ 1042.004018] env[61936]: _type = "Task" [ 1042.004018] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.011078] env[61936]: DEBUG oslo_vmware.api [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.103140] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-106fea33-ece5-47f2-8ee0-22fab07d38f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.114473] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a807399c-0037-4eb8-bffd-db4e0ae06c13 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.127033] env[61936]: INFO nova.compute.manager [-] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Took 1.38 seconds to deallocate network for instance. [ 1042.146963] env[61936]: DEBUG nova.compute.manager [req-68d3a539-cc3a-4eba-8654-04212076766c req-98de6e87-6602-43f9-af85-3266634b9516 service nova] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Detach interface failed, port_id=1cdd9a15-01e7-40cf-80fc-24bea0b642d3, reason: Instance 5526cbd5-b1ad-453b-8401-eee7aa356606 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1042.204533] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.266407] env[61936]: DEBUG nova.compute.manager [req-814c19cd-0037-4458-83b0-da0a3280d437 req-0e14fada-7e67-445c-b1d2-ba9beb35de6f service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-vif-plugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1042.266407] env[61936]: DEBUG oslo_concurrency.lockutils [req-814c19cd-0037-4458-83b0-da0a3280d437 req-0e14fada-7e67-445c-b1d2-ba9beb35de6f service nova] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.266407] env[61936]: DEBUG oslo_concurrency.lockutils [req-814c19cd-0037-4458-83b0-da0a3280d437 req-0e14fada-7e67-445c-b1d2-ba9beb35de6f service nova] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.266407] env[61936]: DEBUG oslo_concurrency.lockutils [req-814c19cd-0037-4458-83b0-da0a3280d437 req-0e14fada-7e67-445c-b1d2-ba9beb35de6f service nova] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.266407] env[61936]: DEBUG nova.compute.manager [req-814c19cd-0037-4458-83b0-da0a3280d437 req-0e14fada-7e67-445c-b1d2-ba9beb35de6f service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] No waiting events found dispatching network-vif-plugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1042.266407] env[61936]: WARNING nova.compute.manager [req-814c19cd-0037-4458-83b0-da0a3280d437 req-0e14fada-7e67-445c-b1d2-ba9beb35de6f service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received unexpected event network-vif-plugged-d169ca50-5a14-4a65-9da0-ea79c80f75c8 for instance with vm_state shelved_offloaded and task_state spawning. [ 1042.313557] env[61936]: DEBUG nova.network.neutron [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.347832] env[61936]: DEBUG nova.scheduler.client.report [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1042.351745] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.351937] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.352110] env[61936]: DEBUG nova.network.neutron [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1042.510905] env[61936]: DEBUG oslo_vmware.api [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193821} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.511227] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.511462] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1042.511644] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1042.511826] env[61936]: INFO nova.compute.manager [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1042.512095] env[61936]: DEBUG oslo.service.loopingcall [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.512305] env[61936]: DEBUG nova.compute.manager [-] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1042.512397] env[61936]: DEBUG nova.network.neutron [-] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1042.514108] env[61936]: DEBUG nova.network.neutron [-] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.517715] env[61936]: DEBUG nova.network.neutron [-] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.634187] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.705255] env[61936]: DEBUG oslo_vmware.api [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253415, 'name': ReconfigVM_Task, 'duration_secs': 5.792568} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.705686] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.705785] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Reconfigured VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1042.816277] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.854096] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.802s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.856436] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.220s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.858028] env[61936]: INFO nova.compute.claims [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.882450] env[61936]: INFO nova.scheduler.client.report [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted allocations for instance 53605e96-a4d1-4a7b-8fef-1f1878b099c3 [ 1043.017097] env[61936]: INFO nova.compute.manager [-] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Took 1.40 seconds to deallocate network for instance. [ 1043.019391] env[61936]: INFO nova.compute.manager [-] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Took 1.37 seconds to deallocate network for instance. [ 1043.098721] env[61936]: DEBUG nova.network.neutron [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [{"id": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "address": "fa:16:3e:65:44:c9", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd169ca50-5a", "ovs_interfaceid": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.270894] env[61936]: DEBUG nova.network.neutron [-] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.391221] env[61936]: DEBUG oslo_concurrency.lockutils [None req-177079e1-d19e-472a-b3d2-2c73203861f1 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "53605e96-a4d1-4a7b-8fef-1f1878b099c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.598s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.525467] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.531458] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.601283] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.659551] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='40f5a4d0ae3c0df87f204e1c49de2093',container_format='bare',created_at=2024-10-10T16:51:05Z,direct_url=,disk_format='vmdk',id=baa3a86e-79f6-4c48-b14b-11d628f6dec2,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-717127375-shelved',owner='905821a1ff2b4011994c0d2d8bc08b13',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-10T16:51:20Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1043.659832] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.659992] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1043.660200] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.660353] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1043.660504] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1043.660714] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1043.660900] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1043.661322] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1043.661322] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1043.661488] env[61936]: DEBUG nova.virt.hardware [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1043.662402] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e6be4d-a7d8-4d93-b103-95f2f83748f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.667909] env[61936]: DEBUG nova.compute.manager [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Received event network-vif-deleted-a424508f-87c8-48b0-b5de-4a4916cddc3f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1043.668130] env[61936]: DEBUG nova.compute.manager [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Received event network-vif-deleted-7775677e-cf35-458b-8c94-da7dd8dc652a {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1043.668315] env[61936]: DEBUG nova.compute.manager [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-vif-deleted-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1043.668477] env[61936]: INFO nova.compute.manager [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Neutron deleted interface 3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618; detaching it from the instance and deleting it from the info cache [ 1043.668755] env[61936]: DEBUG nova.network.neutron [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fa637c42-69ec-480c-bc7b-08a8a723660b", "address": "fa:16:3e:d0:9d:1d", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa637c42-69", "ovs_interfaceid": "fa637c42-69ec-480c-bc7b-08a8a723660b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.675608] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35bb773-d4d9-4963-a904-7f3ff2e52b7b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.689608] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:44:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd169ca50-5a14-4a65-9da0-ea79c80f75c8', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1043.696967] env[61936]: DEBUG oslo.service.loopingcall [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.697698] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1043.697923] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-911270b1-8bf7-415a-8f86-bc67007918d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.718715] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1043.718715] env[61936]: value = "task-1253438" [ 1043.718715] env[61936]: _type = "Task" [ 1043.718715] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.726561] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253438, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.774227] env[61936]: INFO nova.compute.manager [-] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Took 1.26 seconds to deallocate network for instance. [ 1043.945705] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.945917] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.946110] env[61936]: DEBUG nova.network.neutron [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.062172] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72eedac-b077-49c4-8911-79ccee6a6890 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.076575] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e24488d-45ff-48ac-a41d-59d2ea88204f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.110465] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2949db0d-bf11-4ecb-93aa-7fec94be1015 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.118218] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5721bc0-faad-4d65-8aa5-1060f0196423 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.132997] env[61936]: DEBUG nova.compute.provider_tree [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.171572] env[61936]: DEBUG oslo_concurrency.lockutils [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.171767] env[61936]: DEBUG oslo_concurrency.lockutils [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] Acquired lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.172710] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae69509f-f58d-437e-a8b3-3eca3ffd0cf4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.190612] env[61936]: DEBUG oslo_concurrency.lockutils [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] Releasing lock "0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.190894] env[61936]: WARNING nova.compute.manager [req-f0512d93-bf59-41eb-8e0f-a81ce02b3ee6 req-9a1621d9-9cd1-4675-96dd-99fffee53846 service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Detach interface failed, port_id=3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618, reason: No device with interface-id 3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 exists on VM: nova.exception.NotFound: No device with interface-id 3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618 exists on VM [ 1044.228835] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253438, 'name': CreateVM_Task, 'duration_secs': 0.409841} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.229025] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1044.229752] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.229965] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.230351] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.230610] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-295e795a-7f23-4ce3-a163-481e6d4b075d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.235828] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1044.235828] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52603b8d-7cd9-d942-87c4-03bbc21c25c4" [ 1044.235828] env[61936]: _type = "Task" [ 1044.235828] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.244607] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52603b8d-7cd9-d942-87c4-03bbc21c25c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.281709] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.298306] env[61936]: DEBUG nova.compute.manager [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1044.298633] env[61936]: DEBUG nova.compute.manager [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing instance network info cache due to event network-changed-d169ca50-5a14-4a65-9da0-ea79c80f75c8. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1044.298948] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] Acquiring lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.299199] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] Acquired lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.299468] env[61936]: DEBUG nova.network.neutron [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Refreshing network info cache for port d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1044.353528] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1044.353856] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f15b427a-15b9-488f-91d2-e76c619b77f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.361455] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1044.361455] env[61936]: value = "task-1253439" [ 1044.361455] env[61936]: _type = "Task" [ 1044.361455] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.370411] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.403566] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.403904] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "0137305a-dace-4eda-9d90-7233c001176f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.404170] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "0137305a-dace-4eda-9d90-7233c001176f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.404401] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.404624] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "0137305a-dace-4eda-9d90-7233c001176f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.407085] env[61936]: INFO nova.compute.manager [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Terminating instance [ 1044.636341] env[61936]: DEBUG nova.scheduler.client.report [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1044.686645] env[61936]: INFO nova.network.neutron [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Port fa637c42-69ec-480c-bc7b-08a8a723660b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1044.687047] env[61936]: DEBUG nova.network.neutron [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.746767] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.747051] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Processing image baa3a86e-79f6-4c48-b14b-11d628f6dec2 {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.747297] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2/baa3a86e-79f6-4c48-b14b-11d628f6dec2.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.747449] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2/baa3a86e-79f6-4c48-b14b-11d628f6dec2.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.747627] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.747897] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c823efb-c553-4d95-907d-68335be77a9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.755475] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.755653] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.756338] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c68d816f-a334-496d-84f4-102745f33001 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.761179] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1044.761179] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5260fe5d-f0b9-4e45-61f7-e5d0ddec7130" [ 1044.761179] env[61936]: _type = "Task" [ 1044.761179] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.768244] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5260fe5d-f0b9-4e45-61f7-e5d0ddec7130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.877623] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253439, 'name': PowerOffVM_Task, 'duration_secs': 0.227211} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.877909] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "bbb98011-2321-4a69-a882-a4d54df132cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.878142] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "bbb98011-2321-4a69-a882-a4d54df132cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.879220] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1044.880537] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb9591b-5850-41c5-be85-c3bf8c1fe434 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.902025] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b2035a-4022-4ba8-8b02-c5e4eae3bf0d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.913609] env[61936]: DEBUG nova.compute.manager [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1044.913837] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1044.915150] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b902e75a-a781-43b2-af6d-d8d7463b295e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.922240] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1044.922503] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76d02bd4-2f7b-4431-adea-a0be846c716f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.930585] env[61936]: DEBUG oslo_vmware.api [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1044.930585] env[61936]: value = "task-1253440" [ 1044.930585] env[61936]: _type = "Task" [ 1044.930585] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.936913] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1044.937250] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cf1a2eb-c61e-4752-a369-238efb871798 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.943978] env[61936]: DEBUG oslo_vmware.api [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253440, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.945242] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1044.945242] env[61936]: value = "task-1253441" [ 1044.945242] env[61936]: _type = "Task" [ 1044.945242] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.953353] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1044.953629] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.953800] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.953951] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.954152] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.954400] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de5d0592-3fbe-465a-a8da-0455c79dd9db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.961478] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.961657] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.962431] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67442d3c-ebb6-45a2-9242-4bed28a547da {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.966909] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1044.966909] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527eadda-9b4f-7831-08ff-c1289224bced" [ 1044.966909] env[61936]: _type = "Task" [ 1044.966909] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.974563] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527eadda-9b4f-7831-08ff-c1289224bced, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.080424] env[61936]: DEBUG nova.network.neutron [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updated VIF entry in instance network info cache for port d169ca50-5a14-4a65-9da0-ea79c80f75c8. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.080964] env[61936]: DEBUG nova.network.neutron [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [{"id": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "address": "fa:16:3e:65:44:c9", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd169ca50-5a", "ovs_interfaceid": "d169ca50-5a14-4a65-9da0-ea79c80f75c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.143388] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.143940] env[61936]: DEBUG nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1045.147102] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.513s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.147160] env[61936]: DEBUG nova.objects.instance [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lazy-loading 'resources' on Instance uuid 5526cbd5-b1ad-453b-8401-eee7aa356606 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.189788] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-0137305a-dace-4eda-9d90-7233c001176f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.272450] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Preparing fetch location {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1045.272735] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Fetch image to [datastore2] OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a/OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a.vmdk {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1045.272927] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Downloading stream optimized image baa3a86e-79f6-4c48-b14b-11d628f6dec2 to [datastore2] OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a/OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a.vmdk on the data store datastore2 as vApp {{(pid=61936) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1045.273123] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Downloading image file data baa3a86e-79f6-4c48-b14b-11d628f6dec2 to the ESX as VM named 'OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a' {{(pid=61936) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1045.343916] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1045.343916] env[61936]: value = "resgroup-9" [ 1045.343916] env[61936]: _type = "ResourcePool" [ 1045.343916] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1045.344298] env[61936]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-01f78083-3023-44e2-b2ec-3d7cc5851384 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.367148] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease: (returnval){ [ 1045.367148] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528b8e2a-a968-f68e-dc9e-1cfcff3fc52f" [ 1045.367148] env[61936]: _type = "HttpNfcLease" [ 1045.367148] env[61936]: } obtained for vApp import into resource pool (val){ [ 1045.367148] env[61936]: value = "resgroup-9" [ 1045.367148] env[61936]: _type = "ResourcePool" [ 1045.367148] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1045.367407] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the lease: (returnval){ [ 1045.367407] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528b8e2a-a968-f68e-dc9e-1cfcff3fc52f" [ 1045.367407] env[61936]: _type = "HttpNfcLease" [ 1045.367407] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1045.373503] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1045.373503] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528b8e2a-a968-f68e-dc9e-1cfcff3fc52f" [ 1045.373503] env[61936]: _type = "HttpNfcLease" [ 1045.373503] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1045.384220] env[61936]: DEBUG nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1045.442287] env[61936]: DEBUG oslo_vmware.api [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253440, 'name': PowerOffVM_Task, 'duration_secs': 0.190453} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.442648] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1045.442871] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1045.443150] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-215c94f6-0316-490b-b364-e8d19198b339 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.476871] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]527eadda-9b4f-7831-08ff-c1289224bced, 'name': SearchDatastore_Task, 'duration_secs': 0.009735} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.477707] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f671324b-9658-43a1-a366-d0dfe81ab04a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.483225] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1045.483225] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5250ecfb-809c-b3c1-1f67-7595f7adc2e0" [ 1045.483225] env[61936]: _type = "Task" [ 1045.483225] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.491070] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5250ecfb-809c-b3c1-1f67-7595f7adc2e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.556589] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1045.556830] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1045.557027] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleting the datastore file [datastore2] 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.557304] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e475eaa-a953-4c5d-9ab5-58fd9a238cc0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.563481] env[61936]: DEBUG oslo_vmware.api [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1045.563481] env[61936]: value = "task-1253444" [ 1045.563481] env[61936]: _type = "Task" [ 1045.563481] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.572393] env[61936]: DEBUG oslo_vmware.api [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.584184] env[61936]: DEBUG oslo_concurrency.lockutils [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] Releasing lock "refresh_cache-fbde880a-b47c-49e8-b84d-0efee21d62c0" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.584506] env[61936]: DEBUG nova.compute.manager [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Received event network-vif-deleted-b56dc652-5710-432a-a37f-d8598b217179 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1045.584763] env[61936]: DEBUG nova.compute.manager [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-vif-deleted-fa637c42-69ec-480c-bc7b-08a8a723660b {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1045.584916] env[61936]: INFO nova.compute.manager [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Neutron deleted interface fa637c42-69ec-480c-bc7b-08a8a723660b; detaching it from the instance and deleting it from the info cache [ 1045.585231] env[61936]: DEBUG nova.network.neutron [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [{"id": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "address": "fa:16:3e:ee:57:e3", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.177", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcaac4e2a-a9", "ovs_interfaceid": "caac4e2a-a9cd-41ca-8d82-9631afe6ed91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.653120] env[61936]: DEBUG nova.compute.utils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1045.654816] env[61936]: DEBUG nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1045.655487] env[61936]: DEBUG nova.network.neutron [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1045.693487] env[61936]: DEBUG oslo_concurrency.lockutils [None req-cfd29455-091a-436e-96fb-37e8a594097e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-0137305a-dace-4eda-9d90-7233c001176f-3ccd46c0-9471-4c4c-aa3f-ae0dc9c46618" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.618s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.712900] env[61936]: DEBUG nova.policy [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '288fd1d0d8e341999481fc62fe14195e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f1b64c32564b9cbb68e3799b456641', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1045.859357] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62f3d17-1c56-4ed2-82c2-a5bb93064914 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.875026] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f94134-92ce-4a09-bc9d-6b4ae4a9652a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.884059] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1045.884059] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528b8e2a-a968-f68e-dc9e-1cfcff3fc52f" [ 1045.884059] env[61936]: _type = "HttpNfcLease" [ 1045.884059] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1045.935956] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7798ffd-a6bd-460e-84ef-063fbf78fc6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.945440] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2916f1b4-19ad-4362-b81b-c748b9a3174b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.950748] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.961103] env[61936]: DEBUG nova.compute.provider_tree [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.982368] env[61936]: DEBUG nova.network.neutron [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Successfully created port: 21257d10-73cd-4b46-bc5d-f6f78281e576 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.994236] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5250ecfb-809c-b3c1-1f67-7595f7adc2e0, 'name': SearchDatastore_Task, 'duration_secs': 0.012521} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.994507] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.994761] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 484e5d41-29f4-4845-9633-157c03766978/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk. {{(pid=61936) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1045.995044] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-768279c0-090c-47b6-b059-0db28eed80b6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.001841] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1046.001841] env[61936]: value = "task-1253445" [ 1046.001841] env[61936]: _type = "Task" [ 1046.001841] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.010135] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.072952] env[61936]: DEBUG oslo_vmware.api [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194414} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.073229] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.073412] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1046.073590] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.073762] env[61936]: INFO nova.compute.manager [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1046.074198] env[61936]: DEBUG oslo.service.loopingcall [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.074198] env[61936]: DEBUG nova.compute.manager [-] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1046.074344] env[61936]: DEBUG nova.network.neutron [-] [instance: 0137305a-dace-4eda-9d90-7233c001176f] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1046.088557] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ddb3724-e4a7-4e83-b9b7-a02b1e3841df {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.097035] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e5cfbd-d4be-43c4-b7fe-08b1bf9346c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.125644] env[61936]: DEBUG nova.compute.manager [req-1ffc8b92-eccd-41c4-8732-5d167d1bf642 req-4d30e3cb-9087-46e8-955c-9a0a2622d6df service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Detach interface failed, port_id=fa637c42-69ec-480c-bc7b-08a8a723660b, reason: Instance 0137305a-dace-4eda-9d90-7233c001176f could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1046.126385] env[61936]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port fa637c42-69ec-480c-bc7b-08a8a723660b could not be found.", "detail": ""}} {{(pid=61936) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1046.126598] env[61936]: DEBUG nova.network.neutron [-] Unable to show port fa637c42-69ec-480c-bc7b-08a8a723660b as it no longer exists. {{(pid=61936) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1046.160711] env[61936]: DEBUG nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1046.379387] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1046.379387] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528b8e2a-a968-f68e-dc9e-1cfcff3fc52f" [ 1046.379387] env[61936]: _type = "HttpNfcLease" [ 1046.379387] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1046.379899] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1046.379899] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528b8e2a-a968-f68e-dc9e-1cfcff3fc52f" [ 1046.379899] env[61936]: _type = "HttpNfcLease" [ 1046.379899] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1046.380697] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06af57c-3b16-47cd-8392-489373857248 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.391538] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526581ee-c423-faf7-2f13-a04b3612d191/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1046.392075] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526581ee-c423-faf7-2f13-a04b3612d191/disk-0.vmdk. {{(pid=61936) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1046.467462] env[61936]: DEBUG nova.scheduler.client.report [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1046.481938] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-751ec688-65bb-4bcb-baed-99da7ccb1343 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.513735] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253445, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.806671] env[61936]: DEBUG nova.compute.manager [req-e266c7ac-40a0-4830-85e5-1bac8d8043ec req-a17f2dbf-2c39-4b9d-a920-62ff61cf5bbb service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Received event network-vif-deleted-caac4e2a-a9cd-41ca-8d82-9631afe6ed91 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1046.806909] env[61936]: INFO nova.compute.manager [req-e266c7ac-40a0-4830-85e5-1bac8d8043ec req-a17f2dbf-2c39-4b9d-a920-62ff61cf5bbb service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Neutron deleted interface caac4e2a-a9cd-41ca-8d82-9631afe6ed91; detaching it from the instance and deleting it from the info cache [ 1046.807245] env[61936]: DEBUG nova.network.neutron [req-e266c7ac-40a0-4830-85e5-1bac8d8043ec req-a17f2dbf-2c39-4b9d-a920-62ff61cf5bbb service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.977533] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.829s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.978913] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.454s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.979448] env[61936]: DEBUG nova.objects.instance [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lazy-loading 'resources' on Instance uuid 1c391c45-a041-422d-9a44-a29306f99a6d {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.004859] env[61936]: INFO nova.scheduler.client.report [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Deleted allocations for instance 5526cbd5-b1ad-453b-8401-eee7aa356606 [ 1047.020398] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591372} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.022621] env[61936]: INFO nova.virt.vmwareapi.ds_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 484e5d41-29f4-4845-9633-157c03766978/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk. [ 1047.023477] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589fbc23-b402-4f9f-9b94-2d80b2907012 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.054124] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 484e5d41-29f4-4845-9633-157c03766978/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1047.060157] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5246526d-6f50-4cb4-b45f-5def647d910c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.085122] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1047.085122] env[61936]: value = "task-1253446" [ 1047.085122] env[61936]: _type = "Task" [ 1047.085122] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.094579] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253446, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.172561] env[61936]: DEBUG nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1047.200963] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1047.201275] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1047.201436] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1047.201673] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1047.201760] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1047.201926] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1047.202149] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1047.202347] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1047.202534] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1047.202788] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1047.202868] env[61936]: DEBUG nova.virt.hardware [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1047.203781] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d223ba5-93d3-4d23-b785-a10f944cc4bc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.214113] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e671a6-f770-49fd-82e8-de6d3728634a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.284587] env[61936]: DEBUG nova.network.neutron [-] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.311192] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1d564590-ede7-4770-84bd-cb4a4163b3da {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.325031] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f2991f-f07e-4900-8ff0-2550f5fb74c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.357663] env[61936]: DEBUG nova.compute.manager [req-e266c7ac-40a0-4830-85e5-1bac8d8043ec req-a17f2dbf-2c39-4b9d-a920-62ff61cf5bbb service nova] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Detach interface failed, port_id=caac4e2a-a9cd-41ca-8d82-9631afe6ed91, reason: Instance 0137305a-dace-4eda-9d90-7233c001176f could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1047.414901] env[61936]: DEBUG nova.compute.manager [req-ea3d317a-057e-4439-a62e-f024ec7861d6 req-cdf24c31-8d1d-4a33-9214-8f420953256c service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Received event network-vif-plugged-21257d10-73cd-4b46-bc5d-f6f78281e576 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1047.415160] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3d317a-057e-4439-a62e-f024ec7861d6 req-cdf24c31-8d1d-4a33-9214-8f420953256c service nova] Acquiring lock "160d7915-3795-460d-961f-7b43c8ed0168-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.415375] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3d317a-057e-4439-a62e-f024ec7861d6 req-cdf24c31-8d1d-4a33-9214-8f420953256c service nova] Lock "160d7915-3795-460d-961f-7b43c8ed0168-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.415565] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3d317a-057e-4439-a62e-f024ec7861d6 req-cdf24c31-8d1d-4a33-9214-8f420953256c service nova] Lock "160d7915-3795-460d-961f-7b43c8ed0168-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.415743] env[61936]: DEBUG nova.compute.manager [req-ea3d317a-057e-4439-a62e-f024ec7861d6 req-cdf24c31-8d1d-4a33-9214-8f420953256c service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] No waiting events found dispatching network-vif-plugged-21257d10-73cd-4b46-bc5d-f6f78281e576 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1047.415980] env[61936]: WARNING nova.compute.manager [req-ea3d317a-057e-4439-a62e-f024ec7861d6 req-cdf24c31-8d1d-4a33-9214-8f420953256c service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Received unexpected event network-vif-plugged-21257d10-73cd-4b46-bc5d-f6f78281e576 for instance with vm_state building and task_state spawning. [ 1047.517834] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ae9bc1f4-97d6-40ee-8327-9c3098a601a4 tempest-ServersNegativeTestJSON-329101335 tempest-ServersNegativeTestJSON-329101335-project-member] Lock "5526cbd5-b1ad-453b-8401-eee7aa356606" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.479s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.539083] env[61936]: DEBUG nova.network.neutron [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Successfully updated port: 21257d10-73cd-4b46-bc5d-f6f78281e576 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1047.604489] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253446, 'name': ReconfigVM_Task, 'duration_secs': 0.361441} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.604959] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 484e5d41-29f4-4845-9633-157c03766978/43acc3d3-5e18-42a0-9168-cb6831c6bbfb-rescue.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.606118] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d7f4ce-9f87-4d42-a4a0-48896f4955c8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.640597] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-713a5527-0967-493b-8faa-12fa8f6c1a3a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.656560] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1047.656560] env[61936]: value = "task-1253447" [ 1047.656560] env[61936]: _type = "Task" [ 1047.656560] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.666131] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253447, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.743206] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387e8e06-7d19-4a42-b5bf-5770c747d16d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.751687] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b24f149-c2c3-49a6-9435-70ff7cc0740e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.789434] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Completed reading data from the image iterator. {{(pid=61936) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1047.789710] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526581ee-c423-faf7-2f13-a04b3612d191/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1047.790568] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f37259-704e-4bb1-a18f-febf2e4ca823 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.793602] env[61936]: INFO nova.compute.manager [-] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Took 1.72 seconds to deallocate network for instance. [ 1047.794375] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d6fd5b-0460-49f9-b2e6-22363387c35b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.805470] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526581ee-c423-faf7-2f13-a04b3612d191/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1047.805665] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526581ee-c423-faf7-2f13-a04b3612d191/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1047.807845] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b1d6af39-62e8-4123-8433-3509670a78d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.810782] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1742885a-2d2e-4b17-a8a0-f0bd513c2109 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.825279] env[61936]: DEBUG nova.compute.provider_tree [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.002618] env[61936]: DEBUG oslo_vmware.rw_handles [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526581ee-c423-faf7-2f13-a04b3612d191/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1048.002889] env[61936]: INFO nova.virt.vmwareapi.images [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Downloaded image file data baa3a86e-79f6-4c48-b14b-11d628f6dec2 [ 1048.003765] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396b0618-2ddd-476b-a7b7-19e4c52eeaa3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.019452] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9357a01b-9b80-4fa7-b770-6482229e46aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.046308] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "refresh_cache-160d7915-3795-460d-961f-7b43c8ed0168" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.046308] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "refresh_cache-160d7915-3795-460d-961f-7b43c8ed0168" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.046308] env[61936]: DEBUG nova.network.neutron [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1048.048266] env[61936]: INFO nova.virt.vmwareapi.images [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] The imported VM was unregistered [ 1048.050729] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Caching image {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1048.050989] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating directory with path [datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2 {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.051431] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c44e217-1c57-40fc-8111-97d38751d0f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.063706] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created directory with path [datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2 {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.063910] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a/OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a.vmdk to [datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2/baa3a86e-79f6-4c48-b14b-11d628f6dec2.vmdk. {{(pid=61936) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1048.064181] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-75b2bf0a-008c-4520-8619-410b387f7ab0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.071098] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1048.071098] env[61936]: value = "task-1253449" [ 1048.071098] env[61936]: _type = "Task" [ 1048.071098] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.079221] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.169309] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253447, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.303531] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.331032] env[61936]: DEBUG nova.scheduler.client.report [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1048.584878] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.586691] env[61936]: DEBUG nova.network.neutron [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1048.670635] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253447, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.729476] env[61936]: DEBUG nova.network.neutron [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Updating instance_info_cache with network_info: [{"id": "21257d10-73cd-4b46-bc5d-f6f78281e576", "address": "fa:16:3e:b6:2d:6d", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21257d10-73", "ovs_interfaceid": "21257d10-73cd-4b46-bc5d-f6f78281e576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.834693] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.837206] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.306s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.837471] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.839365] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.558s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.839595] env[61936]: DEBUG nova.objects.instance [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lazy-loading 'resources' on Instance uuid b4fb4cfc-10a5-4b24-9f79-1e6832659f89 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.859110] env[61936]: INFO nova.scheduler.client.report [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Deleted allocations for instance 1c391c45-a041-422d-9a44-a29306f99a6d [ 1048.864011] env[61936]: INFO nova.scheduler.client.report [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted allocations for instance 755c5fea-e3d2-4cb3-b717-5f862401d7e1 [ 1049.082920] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.173021] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253447, 'name': ReconfigVM_Task, 'duration_secs': 1.213696} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.173021] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.173021] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56e79337-b9d1-432d-ab0a-23826f9ef1eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.183592] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1049.183592] env[61936]: value = "task-1253450" [ 1049.183592] env[61936]: _type = "Task" [ 1049.183592] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.192083] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.233030] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "refresh_cache-160d7915-3795-460d-961f-7b43c8ed0168" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.233546] env[61936]: DEBUG nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Instance network_info: |[{"id": "21257d10-73cd-4b46-bc5d-f6f78281e576", "address": "fa:16:3e:b6:2d:6d", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21257d10-73", "ovs_interfaceid": "21257d10-73cd-4b46-bc5d-f6f78281e576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1049.234012] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:2d:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b94712a6-b777-47dd-bc06-f9acfce2d936', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21257d10-73cd-4b46-bc5d-f6f78281e576', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.244087] env[61936]: DEBUG oslo.service.loopingcall [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.244451] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1049.244705] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af0b10b3-d601-433b-b73c-c3485f720f21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.268327] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.268327] env[61936]: value = "task-1253451" [ 1049.268327] env[61936]: _type = "Task" [ 1049.268327] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.279733] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253451, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.368954] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f45c8fb4-2201-469c-a89b-fc3b1f0dabdd tempest-ImagesTestJSON-578844145 tempest-ImagesTestJSON-578844145-project-member] Lock "1c391c45-a041-422d-9a44-a29306f99a6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.389s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.373911] env[61936]: DEBUG oslo_concurrency.lockutils [None req-be280236-3bb8-4e87-a317-e38cce914694 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "755c5fea-e3d2-4cb3-b717-5f862401d7e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.358s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.472362] env[61936]: DEBUG nova.compute.manager [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Received event network-changed-21257d10-73cd-4b46-bc5d-f6f78281e576 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1049.472707] env[61936]: DEBUG nova.compute.manager [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Refreshing instance network info cache due to event network-changed-21257d10-73cd-4b46-bc5d-f6f78281e576. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1049.473170] env[61936]: DEBUG oslo_concurrency.lockutils [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] Acquiring lock "refresh_cache-160d7915-3795-460d-961f-7b43c8ed0168" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.473400] env[61936]: DEBUG oslo_concurrency.lockutils [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] Acquired lock "refresh_cache-160d7915-3795-460d-961f-7b43c8ed0168" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.473631] env[61936]: DEBUG nova.network.neutron [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Refreshing network info cache for port 21257d10-73cd-4b46-bc5d-f6f78281e576 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1049.536760] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adef6448-f0a9-462d-85d2-279379da1ce9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.550765] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f07aabe-4d09-4948-a7fd-2d8b54574d78 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.602508] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9fc4b99-99f8-4262-b1a9-7dd43fe6bac3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.614942] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.616560] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0236250f-7703-4861-ac12-37fcbd3f6c62 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.635817] env[61936]: DEBUG nova.compute.provider_tree [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.695184] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253450, 'name': PowerOnVM_Task} progress is 92%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.780998] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253451, 'name': CreateVM_Task, 'duration_secs': 0.390406} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.781258] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1049.782027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.782209] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.782547] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1049.782814] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2e7dc9c-d14e-4961-acf9-2da213143163 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.795153] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1049.795153] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a084fd-fe96-786e-a5fb-73474d83316f" [ 1049.795153] env[61936]: _type = "Task" [ 1049.795153] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.806277] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a084fd-fe96-786e-a5fb-73474d83316f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.110405] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.139239] env[61936]: DEBUG nova.scheduler.client.report [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1050.196784] env[61936]: DEBUG oslo_vmware.api [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253450, 'name': PowerOnVM_Task, 'duration_secs': 0.561406} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.197174] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.200108] env[61936]: DEBUG nova.compute.manager [None req-9a0b39d8-f0e4-451f-8216-d9774aba906f tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1050.201011] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a3a8fa-af27-41c4-901b-c2051294948d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.226680] env[61936]: DEBUG nova.network.neutron [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Updated VIF entry in instance network info cache for port 21257d10-73cd-4b46-bc5d-f6f78281e576. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1050.227079] env[61936]: DEBUG nova.network.neutron [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Updating instance_info_cache with network_info: [{"id": "21257d10-73cd-4b46-bc5d-f6f78281e576", "address": "fa:16:3e:b6:2d:6d", "network": {"id": "8d761efd-18e6-41dc-b278-56e6c074ba29", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-412996895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d4f1b64c32564b9cbb68e3799b456641", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b94712a6-b777-47dd-bc06-f9acfce2d936", "external-id": "nsx-vlan-transportzone-494", "segmentation_id": 494, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21257d10-73", "ovs_interfaceid": "21257d10-73cd-4b46-bc5d-f6f78281e576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.308150] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a084fd-fe96-786e-a5fb-73474d83316f, 'name': SearchDatastore_Task, 'duration_secs': 0.093395} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.308477] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.308719] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.308962] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.309133] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.309321] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.309598] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45d9cf5e-94d9-4d53-b5f1-10538a3f1600 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.330944] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.331144] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1050.331920] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-437cf63b-8c1a-4b79-8087-173b96789687 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.340087] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1050.340087] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e4a095-d0f4-f3f6-f25a-4e32ba4c79b6" [ 1050.340087] env[61936]: _type = "Task" [ 1050.340087] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.350350] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e4a095-d0f4-f3f6-f25a-4e32ba4c79b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.610360] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.648597] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.809s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.651834] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.700s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.652782] env[61936]: INFO nova.compute.claims [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.687519] env[61936]: INFO nova.scheduler.client.report [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted allocations for instance b4fb4cfc-10a5-4b24-9f79-1e6832659f89 [ 1050.732866] env[61936]: DEBUG oslo_concurrency.lockutils [req-3ecedf37-a9d4-4398-bc60-729317eccb76 req-e1390599-736c-4783-a829-5f79cd706b41 service nova] Releasing lock "refresh_cache-160d7915-3795-460d-961f-7b43c8ed0168" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.852634] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e4a095-d0f4-f3f6-f25a-4e32ba4c79b6, 'name': SearchDatastore_Task, 'duration_secs': 0.082069} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.853658] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6ce2e08-2a46-45a2-a520-bcaa66cea17f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.861185] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1050.861185] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ae1a2c-0ead-1cd9-e634-cf4f0840d5ec" [ 1050.861185] env[61936]: _type = "Task" [ 1050.861185] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.871201] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ae1a2c-0ead-1cd9-e634-cf4f0840d5ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.111682] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.200180] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d40fc13-28c0-412c-8bcd-6be8dff8da89 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "b4fb4cfc-10a5-4b24-9f79-1e6832659f89" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.316s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.371341] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ae1a2c-0ead-1cd9-e634-cf4f0840d5ec, 'name': SearchDatastore_Task, 'duration_secs': 0.086361} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.371606] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.371867] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 160d7915-3795-460d-961f-7b43c8ed0168/160d7915-3795-460d-961f-7b43c8ed0168.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1051.372145] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-295add97-80eb-494a-9f8e-41f295afdc8d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.378828] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1051.378828] env[61936]: value = "task-1253452" [ 1051.378828] env[61936]: _type = "Task" [ 1051.378828] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.386332] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.457270] env[61936]: INFO nova.compute.manager [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Unrescuing [ 1051.457538] env[61936]: DEBUG oslo_concurrency.lockutils [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.457692] env[61936]: DEBUG oslo_concurrency.lockutils [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquired lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.457869] env[61936]: DEBUG nova.network.neutron [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1051.612140] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.790108] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1cd279-f1d4-4adf-b63d-40ce1335b914 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.798647] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8782736a-1ff9-4bc2-9fbe-807263d696af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.829482] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926ae7bf-71a7-462c-a9e3-d879e95d55f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.837826] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e7dcb5-9466-43c5-996a-a31436c0e3a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.851627] env[61936]: DEBUG nova.compute.provider_tree [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.892098] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253452, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.113304] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253449, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.545356} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.113612] env[61936]: INFO nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a/OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a.vmdk to [datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2/baa3a86e-79f6-4c48-b14b-11d628f6dec2.vmdk. [ 1052.113817] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Cleaning up location [datastore2] OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1052.114026] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_ad6baa91-f41a-429b-a8bd-5a2ae147649a {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.114349] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9f45c63-d666-4af2-8975-8c7749ec64e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.122442] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1052.122442] env[61936]: value = "task-1253453" [ 1052.122442] env[61936]: _type = "Task" [ 1052.122442] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.135487] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253453, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.223299] env[61936]: DEBUG nova.network.neutron [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.358025] env[61936]: DEBUG nova.scheduler.client.report [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1052.393595] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.846492} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.393595] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 160d7915-3795-460d-961f-7b43c8ed0168/160d7915-3795-460d-961f-7b43c8ed0168.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.393595] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.393595] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9434cdbb-430d-4fef-85df-0218f80315e3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.400724] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1052.400724] env[61936]: value = "task-1253454" [ 1052.400724] env[61936]: _type = "Task" [ 1052.400724] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.409828] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253454, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.634030] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253453, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108299} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.634030] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.634030] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2/baa3a86e-79f6-4c48-b14b-11d628f6dec2.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.634030] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2/baa3a86e-79f6-4c48-b14b-11d628f6dec2.vmdk to [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1052.634030] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24248fe9-716d-4d1d-9f41-d47a40858e8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.640275] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1052.640275] env[61936]: value = "task-1253455" [ 1052.640275] env[61936]: _type = "Task" [ 1052.640275] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.648167] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253455, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.728052] env[61936]: DEBUG oslo_concurrency.lockutils [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Releasing lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.728052] env[61936]: DEBUG nova.objects.instance [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lazy-loading 'flavor' on Instance uuid 484e5d41-29f4-4845-9633-157c03766978 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.856648] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "c6043bea-03be-427d-9f39-43f81b0788ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.856876] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.862883] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.212s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.863344] env[61936]: DEBUG nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1052.865610] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.562s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.865829] env[61936]: DEBUG nova.objects.instance [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'resources' on Instance uuid 0137305a-dace-4eda-9d90-7233c001176f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.910970] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253454, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.251682} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.911299] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1052.912087] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52b5545-be11-44ca-a38a-ab6816f9cc07 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.934461] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 160d7915-3795-460d-961f-7b43c8ed0168/160d7915-3795-460d-961f-7b43c8ed0168.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.934773] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69794f6d-4905-4b45-b0ad-6069872843f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.954760] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1052.954760] env[61936]: value = "task-1253456" [ 1052.954760] env[61936]: _type = "Task" [ 1052.954760] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.962764] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253456, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.150437] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253455, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.233469] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a35b00d-13c2-4aac-83ea-0b5c4f379990 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.259885] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.260268] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8563a167-ab87-4a77-a305-763743266d51 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.266320] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1053.266320] env[61936]: value = "task-1253457" [ 1053.266320] env[61936]: _type = "Task" [ 1053.266320] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.274096] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253457, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.359200] env[61936]: DEBUG nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1053.371345] env[61936]: DEBUG nova.compute.utils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.374502] env[61936]: DEBUG nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1053.374502] env[61936]: DEBUG nova.network.neutron [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1053.461316] env[61936]: DEBUG nova.policy [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcee2be71ae044338df4422a3c575ca4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e81b119251964bdb8e5a0ea84b29f2a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1053.469704] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253456, 'name': ReconfigVM_Task, 'duration_secs': 0.315926} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.472842] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 160d7915-3795-460d-961f-7b43c8ed0168/160d7915-3795-460d-961f-7b43c8ed0168.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.474744] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7052723-8666-4bc1-b9bc-d2e91275679f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.482559] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1053.482559] env[61936]: value = "task-1253458" [ 1053.482559] env[61936]: _type = "Task" [ 1053.482559] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.495047] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253458, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.619197] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d809d4ec-ee24-4dcf-8cc7-cf64b2a43eab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.630186] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7453fe-4693-4aae-93bb-1a972392f427 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.670294] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2bd19a3-5dca-4195-9dc3-3b944d4a115e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.683949] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b91a167-0e6e-4fb6-994f-fc8b8ae6feeb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.688071] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253455, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.703250] env[61936]: DEBUG nova.compute.provider_tree [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.778653] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253457, 'name': PowerOffVM_Task, 'duration_secs': 0.317801} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.780177] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.785620] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfiguring VM instance instance-0000005b to detach disk 2002 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1053.785795] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c9a1676-756b-43fa-bdc7-f74d28bfd905 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.805787] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1053.805787] env[61936]: value = "task-1253459" [ 1053.805787] env[61936]: _type = "Task" [ 1053.805787] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.817453] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253459, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.879073] env[61936]: DEBUG nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1053.888382] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.988104] env[61936]: DEBUG nova.network.neutron [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Successfully created port: 988c6387-6b32-4fc0-a731-70a6a97bc815 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.997997] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253458, 'name': Rename_Task, 'duration_secs': 0.15819} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.998364] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1053.998634] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f199cf7d-4aa7-4211-ad2b-df094c6b864e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.007659] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1054.007659] env[61936]: value = "task-1253460" [ 1054.007659] env[61936]: _type = "Task" [ 1054.007659] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.017893] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253460, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.182184] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253455, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.206217] env[61936]: DEBUG nova.scheduler.client.report [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1054.324996] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253459, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.521356] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253460, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.678736] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253455, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.716590] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.720341] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.832s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.722050] env[61936]: INFO nova.compute.claims [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.752888] env[61936]: INFO nova.scheduler.client.report [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted allocations for instance 0137305a-dace-4eda-9d90-7233c001176f [ 1054.819599] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253459, 'name': ReconfigVM_Task, 'duration_secs': 0.527991} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.819915] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfigured VM instance instance-0000005b to detach disk 2002 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1054.820147] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.820428] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5a46407-d518-4c51-96ae-a77585092f3a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.829544] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1054.829544] env[61936]: value = "task-1253461" [ 1054.829544] env[61936]: _type = "Task" [ 1054.829544] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.841864] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253461, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.889220] env[61936]: DEBUG nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1054.917843] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1054.918185] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.918359] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1054.918546] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.918695] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1054.918842] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1054.919074] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1054.919251] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1054.919419] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1054.919584] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1054.919758] env[61936]: DEBUG nova.virt.hardware [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1054.921103] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-138fa783-8fba-4083-8079-c46899b79256 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.931594] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505c3402-ff10-443b-acef-f3476f32ccdf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.022017] env[61936]: DEBUG oslo_vmware.api [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253460, 'name': PowerOnVM_Task, 'duration_secs': 0.598748} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.022397] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.022646] env[61936]: INFO nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Took 7.85 seconds to spawn the instance on the hypervisor. [ 1055.022879] env[61936]: DEBUG nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1055.024163] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a13e59a-cef9-4cac-9575-e096a66fb7f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.180721] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253455, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.263442] env[61936]: DEBUG oslo_concurrency.lockutils [None req-839ab5fb-6d08-4978-ad28-e931fcb21d2e tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "0137305a-dace-4eda-9d90-7233c001176f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.859s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.342035] env[61936]: DEBUG oslo_vmware.api [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253461, 'name': PowerOnVM_Task, 'duration_secs': 0.509972} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.342035] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.342402] env[61936]: DEBUG nova.compute.manager [None req-119f4816-7db0-481e-b0eb-2db52205318a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1055.343431] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e207bc5d-8b2f-4074-9b9d-496586d6ae8e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.546788] env[61936]: INFO nova.compute.manager [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Took 13.93 seconds to build instance. [ 1055.679110] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253455, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.952841} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.679630] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/baa3a86e-79f6-4c48-b14b-11d628f6dec2/baa3a86e-79f6-4c48-b14b-11d628f6dec2.vmdk to [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1055.680476] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f23cce3-5d55-4a51-b448-d3d1db720d8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.706070] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1055.706434] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-726303f7-2b18-4f0b-9c5b-017994207474 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.727139] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1055.727139] env[61936]: value = "task-1253462" [ 1055.727139] env[61936]: _type = "Task" [ 1055.727139] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.740363] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253462, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.745214] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.745457] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.871866] env[61936]: DEBUG nova.compute.manager [req-c717703a-d5a6-456e-b799-a2cb22a700b2 req-edee3dfe-1f1a-41d5-a187-283ff25902af service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Received event network-vif-plugged-988c6387-6b32-4fc0-a731-70a6a97bc815 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1055.872101] env[61936]: DEBUG oslo_concurrency.lockutils [req-c717703a-d5a6-456e-b799-a2cb22a700b2 req-edee3dfe-1f1a-41d5-a187-283ff25902af service nova] Acquiring lock "bbb98011-2321-4a69-a882-a4d54df132cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.872354] env[61936]: DEBUG oslo_concurrency.lockutils [req-c717703a-d5a6-456e-b799-a2cb22a700b2 req-edee3dfe-1f1a-41d5-a187-283ff25902af service nova] Lock "bbb98011-2321-4a69-a882-a4d54df132cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.873321] env[61936]: DEBUG oslo_concurrency.lockutils [req-c717703a-d5a6-456e-b799-a2cb22a700b2 req-edee3dfe-1f1a-41d5-a187-283ff25902af service nova] Lock "bbb98011-2321-4a69-a882-a4d54df132cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.873321] env[61936]: DEBUG nova.compute.manager [req-c717703a-d5a6-456e-b799-a2cb22a700b2 req-edee3dfe-1f1a-41d5-a187-283ff25902af service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] No waiting events found dispatching network-vif-plugged-988c6387-6b32-4fc0-a731-70a6a97bc815 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1055.873321] env[61936]: WARNING nova.compute.manager [req-c717703a-d5a6-456e-b799-a2cb22a700b2 req-edee3dfe-1f1a-41d5-a187-283ff25902af service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Received unexpected event network-vif-plugged-988c6387-6b32-4fc0-a731-70a6a97bc815 for instance with vm_state building and task_state spawning. [ 1055.970229] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df5a134-32b7-4093-97ef-8ddcd5b896e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.979498] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38044c24-b6fa-4836-b360-93b85163170d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.018891] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172bd7b9-d5c1-4c85-96ab-5eb96c53f128 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.027499] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aeee01a-da6a-4f7f-88d1-2f6360791de1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.041953] env[61936]: DEBUG nova.compute.provider_tree [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.048969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2489c37f-39e2-4c53-8f36-dcdf40e475fc tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "160d7915-3795-460d-961f-7b43c8ed0168" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.448s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.116405] env[61936]: DEBUG nova.network.neutron [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Successfully updated port: 988c6387-6b32-4fc0-a731-70a6a97bc815 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1056.240399] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253462, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.247887] env[61936]: DEBUG nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1056.458969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.459302] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.547968] env[61936]: DEBUG nova.scheduler.client.report [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1056.619882] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-bbb98011-2321-4a69-a882-a4d54df132cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.620794] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-bbb98011-2321-4a69-a882-a4d54df132cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.620794] env[61936]: DEBUG nova.network.neutron [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1056.742545] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253462, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.774300] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.965153] env[61936]: DEBUG nova.compute.utils [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.049612] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.050251] env[61936]: DEBUG nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1057.052937] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.279s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.054377] env[61936]: INFO nova.compute.claims [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.157848] env[61936]: DEBUG nova.compute.manager [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Received event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1057.159547] env[61936]: DEBUG nova.compute.manager [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing instance network info cache due to event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1057.159856] env[61936]: DEBUG oslo_concurrency.lockutils [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] Acquiring lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.160048] env[61936]: DEBUG oslo_concurrency.lockutils [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] Acquired lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.160320] env[61936]: DEBUG nova.network.neutron [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.162175] env[61936]: DEBUG nova.network.neutron [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1057.250750] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253462, 'name': ReconfigVM_Task, 'duration_secs': 1.207702} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.251338] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfigured VM instance instance-00000055 to attach disk [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0/fbde880a-b47c-49e8-b84d-0efee21d62c0.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.256019] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'boot_index': 0, 'guest_format': None, 'encryption_options': None, 'encrypted': False, 'encryption_format': None, 'disk_bus': None, 'device_type': 'disk', 'image_id': '43acc3d3-5e18-42a0-9168-cb6831c6bbfb'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270025', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'name': 'volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fbde880a-b47c-49e8-b84d-0efee21d62c0', 'attached_at': '', 'detached_at': '', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'serial': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e'}, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'disk_bus': None, 'device_type': None, 'attachment_id': 'ffb8c048-58e9-4369-956a-c5e63950f039', 'volume_type': None}], 'swap': None} {{(pid=61936) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1057.256019] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1057.256019] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270025', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'name': 'volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fbde880a-b47c-49e8-b84d-0efee21d62c0', 'attached_at': '', 'detached_at': '', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'serial': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1057.256019] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75b8995-030d-4690-b87d-4f424d80176c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.274362] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebac3afc-39a5-4bd6-aca8-7214503f3a71 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.304337] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e/volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.305035] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f84529b2-51a9-4669-a91b-ea566fa7a4ee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.325725] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1057.325725] env[61936]: value = "task-1253463" [ 1057.325725] env[61936]: _type = "Task" [ 1057.325725] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.333184] env[61936]: DEBUG nova.network.neutron [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Updating instance_info_cache with network_info: [{"id": "988c6387-6b32-4fc0-a731-70a6a97bc815", "address": "fa:16:3e:01:d6:3b", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap988c6387-6b", "ovs_interfaceid": "988c6387-6b32-4fc0-a731-70a6a97bc815", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.335840] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253463, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.467315] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.558854] env[61936]: DEBUG nova.compute.utils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.562134] env[61936]: DEBUG nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1057.562309] env[61936]: DEBUG nova.network.neutron [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1057.610083] env[61936]: DEBUG nova.policy [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b05bdcace008487dba580405eb672ee2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90c29449db6c4beaa7f190d9225cb08a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1057.835916] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-bbb98011-2321-4a69-a882-a4d54df132cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.837132] env[61936]: DEBUG nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Instance network_info: |[{"id": "988c6387-6b32-4fc0-a731-70a6a97bc815", "address": "fa:16:3e:01:d6:3b", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap988c6387-6b", "ovs_interfaceid": "988c6387-6b32-4fc0-a731-70a6a97bc815", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1057.837460] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253463, 'name': ReconfigVM_Task, 'duration_secs': 0.356919} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.837804] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:d6:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '988c6387-6b32-4fc0-a731-70a6a97bc815', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.846532] env[61936]: DEBUG oslo.service.loopingcall [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.846990] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfigured VM instance instance-00000055 to attach disk [datastore1] volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e/volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.853010] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1057.853251] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-723fb6ad-7bb2-442b-89b3-0a8eeb1f3009 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.865996] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-251035b2-1d63-4b78-8063-654fb34e5abb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.886927] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1057.886927] env[61936]: value = "task-1253464" [ 1057.886927] env[61936]: _type = "Task" [ 1057.886927] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.888602] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1057.888602] env[61936]: value = "task-1253465" [ 1057.888602] env[61936]: _type = "Task" [ 1057.888602] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.902964] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253464, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.904016] env[61936]: DEBUG nova.network.neutron [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Successfully created port: 24d5309e-de7c-4a1c-a314-cb7858990b81 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1057.911169] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253465, 'name': CreateVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.036819] env[61936]: DEBUG nova.network.neutron [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updated VIF entry in instance network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.036819] env[61936]: DEBUG nova.network.neutron [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.067185] env[61936]: DEBUG nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1058.103464] env[61936]: DEBUG nova.compute.manager [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Received event network-changed-988c6387-6b32-4fc0-a731-70a6a97bc815 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1058.103687] env[61936]: DEBUG nova.compute.manager [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Refreshing instance network info cache due to event network-changed-988c6387-6b32-4fc0-a731-70a6a97bc815. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1058.106009] env[61936]: DEBUG oslo_concurrency.lockutils [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] Acquiring lock "refresh_cache-bbb98011-2321-4a69-a882-a4d54df132cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.106009] env[61936]: DEBUG oslo_concurrency.lockutils [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] Acquired lock "refresh_cache-bbb98011-2321-4a69-a882-a4d54df132cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.106009] env[61936]: DEBUG nova.network.neutron [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Refreshing network info cache for port 988c6387-6b32-4fc0-a731-70a6a97bc815 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1058.240316] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.240316] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.287858] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958e72c4-0834-4a01-ab63-6bb98992916f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.297124] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd57f75-232f-444c-8fd4-d4bb0bdf1f41 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.331917] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ab54bd-2062-4827-929c-5862ddcae97b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.340275] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b42a3f-1548-4f10-b5c4-987bcce66483 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.354696] env[61936]: DEBUG nova.compute.provider_tree [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.402424] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253464, 'name': ReconfigVM_Task, 'duration_secs': 0.219084} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.405969] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270025', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'name': 'volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fbde880a-b47c-49e8-b84d-0efee21d62c0', 'attached_at': '', 'detached_at': '', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'serial': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1058.406456] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253465, 'name': CreateVM_Task, 'duration_secs': 0.436954} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.406657] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-496bed26-b9f7-4d30-a159-43425a827eef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.408271] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.409216] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.409389] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.409707] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1058.410331] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e365d0f4-d466-432a-bbd2-6dd9804d179b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.414236] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1058.414236] env[61936]: value = "task-1253466" [ 1058.414236] env[61936]: _type = "Task" [ 1058.414236] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.415788] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1058.415788] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523e6105-b5db-5975-f1ba-496dc4ff9470" [ 1058.415788] env[61936]: _type = "Task" [ 1058.415788] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.426412] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253466, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.429593] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523e6105-b5db-5975-f1ba-496dc4ff9470, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.538976] env[61936]: DEBUG oslo_concurrency.lockutils [req-e464975d-f505-45f8-abaa-c74084c1a409 req-c6ba5a0a-0f9f-4019-99b2-6678fabb15cf service nova] Releasing lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.584269] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.584646] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.584981] env[61936]: INFO nova.compute.manager [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Attaching volume 0d436f9e-4fc6-4a67-b16e-9d113e7ae779 to /dev/sdb [ 1058.643286] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302c8b1d-2489-4982-85f8-95f4c5ebcc9f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.650908] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cffb816-85ab-48d3-8511-bd052381f503 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.666172] env[61936]: DEBUG nova.virt.block_device [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updating existing volume attachment record: beb45bbe-b61e-452a-992b-fb227ed5c975 {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1058.743411] env[61936]: DEBUG nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1058.836409] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "160d7915-3795-460d-961f-7b43c8ed0168" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.836688] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "160d7915-3795-460d-961f-7b43c8ed0168" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.836906] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "160d7915-3795-460d-961f-7b43c8ed0168-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.837108] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "160d7915-3795-460d-961f-7b43c8ed0168-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.837310] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "160d7915-3795-460d-961f-7b43c8ed0168-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.841708] env[61936]: INFO nova.compute.manager [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Terminating instance [ 1058.857785] env[61936]: DEBUG nova.scheduler.client.report [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1058.875381] env[61936]: DEBUG nova.network.neutron [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Updated VIF entry in instance network info cache for port 988c6387-6b32-4fc0-a731-70a6a97bc815. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.875731] env[61936]: DEBUG nova.network.neutron [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Updating instance_info_cache with network_info: [{"id": "988c6387-6b32-4fc0-a731-70a6a97bc815", "address": "fa:16:3e:01:d6:3b", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap988c6387-6b", "ovs_interfaceid": "988c6387-6b32-4fc0-a731-70a6a97bc815", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.928106] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253466, 'name': Rename_Task, 'duration_secs': 0.276806} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.932312] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1058.932615] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523e6105-b5db-5975-f1ba-496dc4ff9470, 'name': SearchDatastore_Task, 'duration_secs': 0.023725} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.932838] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c627e643-21e7-4f0b-8867-9f4eebc26828 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.934523] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.934763] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.935010] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.935169] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.935354] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.935604] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44f73349-3030-4f4a-b130-f8a588759aba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.942157] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1058.942157] env[61936]: value = "task-1253468" [ 1058.942157] env[61936]: _type = "Task" [ 1058.942157] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.947751] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.947932] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.951366] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcf20f99-e4ff-4bd2-ad88-32136a3f815f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.953556] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253468, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.956798] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1058.956798] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5216ffef-4b16-fb5a-7734-dfe2621ae589" [ 1058.956798] env[61936]: _type = "Task" [ 1058.956798] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.964866] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5216ffef-4b16-fb5a-7734-dfe2621ae589, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.093451] env[61936]: DEBUG nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1059.119340] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1059.119601] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.119759] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1059.119947] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.120113] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1059.120295] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1059.120510] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1059.120677] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1059.120841] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1059.121014] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1059.121238] env[61936]: DEBUG nova.virt.hardware [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1059.122181] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f06a708-1533-4f01-b88a-992130a63c94 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.130505] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7371158-45cd-4bcd-b599-5c5a25ddd3e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.185199] env[61936]: DEBUG nova.compute.manager [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Received event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1059.185418] env[61936]: DEBUG nova.compute.manager [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing instance network info cache due to event network-changed-7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1059.185638] env[61936]: DEBUG oslo_concurrency.lockutils [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] Acquiring lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.185784] env[61936]: DEBUG oslo_concurrency.lockutils [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] Acquired lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.185950] env[61936]: DEBUG nova.network.neutron [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Refreshing network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.269589] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.345696] env[61936]: DEBUG nova.compute.manager [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1059.345984] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.347015] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c3d161-2e39-4c98-bc06-5c3a1e394f28 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.359531] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.359813] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4247c35-8ed2-4c57-9953-b30d6f60acde {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.362960] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.363911] env[61936]: DEBUG nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1059.368385] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.099s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.370126] env[61936]: INFO nova.compute.claims [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1059.373024] env[61936]: DEBUG oslo_vmware.api [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1059.373024] env[61936]: value = "task-1253469" [ 1059.373024] env[61936]: _type = "Task" [ 1059.373024] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.378339] env[61936]: DEBUG oslo_concurrency.lockutils [req-f994d2d0-6c80-4171-9ddc-9afca3295ffa req-c94574e4-abcc-469e-b0a0-c65dce9b7584 service nova] Releasing lock "refresh_cache-bbb98011-2321-4a69-a882-a4d54df132cc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.382793] env[61936]: DEBUG oslo_vmware.api [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253469, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.453304] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253468, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.467934] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5216ffef-4b16-fb5a-7734-dfe2621ae589, 'name': SearchDatastore_Task, 'duration_secs': 0.011918} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.468821] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d84c2c6-83c9-4026-ba92-a47d8207d752 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.475034] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1059.475034] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5218e11a-7527-2af3-934d-9c41bdb895d4" [ 1059.475034] env[61936]: _type = "Task" [ 1059.475034] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.483263] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5218e11a-7527-2af3-934d-9c41bdb895d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.643222] env[61936]: DEBUG nova.network.neutron [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Successfully updated port: 24d5309e-de7c-4a1c-a314-cb7858990b81 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.869069] env[61936]: DEBUG nova.compute.utils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.869995] env[61936]: DEBUG nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1059.870213] env[61936]: DEBUG nova.network.neutron [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1059.885538] env[61936]: DEBUG oslo_vmware.api [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253469, 'name': PowerOffVM_Task, 'duration_secs': 0.18982} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.885538] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.885538] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.888171] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e004a2d-6ba1-4c7b-94e1-5ac9701d2390 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.922838] env[61936]: DEBUG nova.policy [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f378ef8b8b5f4ae59c0d71dd1661bb59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa564b684410493fa0028fd345048e02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.926863] env[61936]: DEBUG nova.network.neutron [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updated VIF entry in instance network info cache for port 7450de29-76d5-40b0-ae76-a79b3455a9bc. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1059.927143] env[61936]: DEBUG nova.network.neutron [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [{"id": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "address": "fa:16:3e:a3:ea:50", "network": {"id": "f7f55885-239e-4044-82a7-f5e2602f1dbb", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1976147364-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.194", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f1b73bc1dea4df0a995655f3bd50d14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7450de29-76", "ovs_interfaceid": "7450de29-76d5-40b0-ae76-a79b3455a9bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.952194] env[61936]: DEBUG oslo_vmware.api [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253468, 'name': PowerOnVM_Task, 'duration_secs': 0.70439} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.952467] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.974513] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.974811] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.975155] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleting the datastore file [datastore2] 160d7915-3795-460d-961f-7b43c8ed0168 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.975690] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f82a7cf-5f6e-463e-b80d-1d57fd8862f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.994064] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5218e11a-7527-2af3-934d-9c41bdb895d4, 'name': SearchDatastore_Task, 'duration_secs': 0.009847} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.995855] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.996260] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] bbb98011-2321-4a69-a882-a4d54df132cc/bbb98011-2321-4a69-a882-a4d54df132cc.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.996622] env[61936]: DEBUG oslo_vmware.api [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for the task: (returnval){ [ 1059.996622] env[61936]: value = "task-1253471" [ 1059.996622] env[61936]: _type = "Task" [ 1059.996622] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.996820] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ff4fe2b-a390-4872-9d42-76576040f78a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.007747] env[61936]: DEBUG oslo_vmware.api [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253471, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.008949] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1060.008949] env[61936]: value = "task-1253472" [ 1060.008949] env[61936]: _type = "Task" [ 1060.008949] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.016422] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253472, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.054969] env[61936]: DEBUG nova.compute.manager [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1060.056100] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de75290-770f-4ad6-abc4-22268f27d56d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.146800] env[61936]: DEBUG nova.compute.manager [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Received event network-vif-plugged-24d5309e-de7c-4a1c-a314-cb7858990b81 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1060.146800] env[61936]: DEBUG oslo_concurrency.lockutils [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] Acquiring lock "c6043bea-03be-427d-9f39-43f81b0788ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.146800] env[61936]: DEBUG oslo_concurrency.lockutils [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] Lock "c6043bea-03be-427d-9f39-43f81b0788ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.146985] env[61936]: DEBUG oslo_concurrency.lockutils [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] Lock "c6043bea-03be-427d-9f39-43f81b0788ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.147215] env[61936]: DEBUG nova.compute.manager [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] No waiting events found dispatching network-vif-plugged-24d5309e-de7c-4a1c-a314-cb7858990b81 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1060.147420] env[61936]: WARNING nova.compute.manager [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Received unexpected event network-vif-plugged-24d5309e-de7c-4a1c-a314-cb7858990b81 for instance with vm_state building and task_state spawning. [ 1060.147608] env[61936]: DEBUG nova.compute.manager [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Received event network-changed-24d5309e-de7c-4a1c-a314-cb7858990b81 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1060.147777] env[61936]: DEBUG nova.compute.manager [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Refreshing instance network info cache due to event network-changed-24d5309e-de7c-4a1c-a314-cb7858990b81. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1060.147962] env[61936]: DEBUG oslo_concurrency.lockutils [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] Acquiring lock "refresh_cache-c6043bea-03be-427d-9f39-43f81b0788ea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.148180] env[61936]: DEBUG oslo_concurrency.lockutils [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] Acquired lock "refresh_cache-c6043bea-03be-427d-9f39-43f81b0788ea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.148391] env[61936]: DEBUG nova.network.neutron [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Refreshing network info cache for port 24d5309e-de7c-4a1c-a314-cb7858990b81 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1060.149711] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "refresh_cache-c6043bea-03be-427d-9f39-43f81b0788ea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.173583] env[61936]: DEBUG nova.network.neutron [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Successfully created port: 40113568-3584-481c-aeb0-65d87745501c {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.374101] env[61936]: DEBUG nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1060.431697] env[61936]: DEBUG oslo_concurrency.lockutils [req-94967e58-9cfc-4fd2-90d3-e328acd55781 req-e97c3060-bb24-4904-b397-51fff0758a44 service nova] Releasing lock "refresh_cache-484e5d41-29f4-4845-9633-157c03766978" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.511712] env[61936]: DEBUG oslo_vmware.api [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Task: {'id': task-1253471, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25859} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.515276] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.515631] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.515725] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.515905] env[61936]: INFO nova.compute.manager [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1060.516237] env[61936]: DEBUG oslo.service.loopingcall [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.519164] env[61936]: DEBUG nova.compute.manager [-] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1060.519258] env[61936]: DEBUG nova.network.neutron [-] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1060.527571] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253472, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.572943] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1601092-74cc-4fb8-aa5b-0d4110483b74 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.577914] env[61936]: DEBUG oslo_concurrency.lockutils [None req-28f0f4ce-19de-4a34-9c96-87c72c2a2ddf tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.181s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.583315] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eceda1a-8f1a-4e35-aff6-621daad217ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.617983] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8ccf3b-97a3-451d-9aac-b75b846ac7e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.624521] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ee3278-7dfa-49d1-9232-cc03f59218fc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.640204] env[61936]: DEBUG nova.compute.provider_tree [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.682537] env[61936]: DEBUG nova.network.neutron [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1060.801710] env[61936]: DEBUG nova.network.neutron [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.021691] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253472, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588899} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.021994] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] bbb98011-2321-4a69-a882-a4d54df132cc/bbb98011-2321-4a69-a882-a4d54df132cc.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.022230] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.022494] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b74f075-7a06-429f-ac19-408eb9ccb270 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.030776] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1061.030776] env[61936]: value = "task-1253474" [ 1061.030776] env[61936]: _type = "Task" [ 1061.030776] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.037226] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.143433] env[61936]: DEBUG nova.scheduler.client.report [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1061.305801] env[61936]: DEBUG oslo_concurrency.lockutils [req-0f72798e-a4fb-498b-8ec7-9526ecc34c46 req-32bc2524-6ca8-4073-8bfe-aeaec9e5bcb7 service nova] Releasing lock "refresh_cache-c6043bea-03be-427d-9f39-43f81b0788ea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.306300] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "refresh_cache-c6043bea-03be-427d-9f39-43f81b0788ea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.306465] env[61936]: DEBUG nova.network.neutron [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1061.333822] env[61936]: DEBUG nova.network.neutron [-] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.385604] env[61936]: DEBUG nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1061.414524] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1061.414779] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.414940] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1061.415143] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.415297] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1061.415448] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1061.415651] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1061.415813] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1061.415981] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1061.416160] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1061.416337] env[61936]: DEBUG nova.virt.hardware [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1061.417258] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93cd271a-c7e4-4089-a89c-bc6efa042c76 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.427266] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd48d02-4879-44be-99a6-3e2a5910e8d3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.539682] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142267} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.539991] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.540894] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906ecd5e-a5ee-4c6b-b71a-477fc4634fe7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.563178] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] bbb98011-2321-4a69-a882-a4d54df132cc/bbb98011-2321-4a69-a882-a4d54df132cc.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.563524] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81e95fda-a0d9-4f65-aedb-79f467cc2b42 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.582951] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1061.582951] env[61936]: value = "task-1253475" [ 1061.582951] env[61936]: _type = "Task" [ 1061.582951] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.591340] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253475, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.648173] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.648757] env[61936]: DEBUG nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1061.715546] env[61936]: DEBUG nova.compute.manager [req-1800eff8-e629-436b-830e-63de334d3ecd req-581b5d6f-8fec-496d-a239-92c4d2d29f75 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Received event network-vif-plugged-40113568-3584-481c-aeb0-65d87745501c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1061.715767] env[61936]: DEBUG oslo_concurrency.lockutils [req-1800eff8-e629-436b-830e-63de334d3ecd req-581b5d6f-8fec-496d-a239-92c4d2d29f75 service nova] Acquiring lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.715980] env[61936]: DEBUG oslo_concurrency.lockutils [req-1800eff8-e629-436b-830e-63de334d3ecd req-581b5d6f-8fec-496d-a239-92c4d2d29f75 service nova] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.716170] env[61936]: DEBUG oslo_concurrency.lockutils [req-1800eff8-e629-436b-830e-63de334d3ecd req-581b5d6f-8fec-496d-a239-92c4d2d29f75 service nova] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.716376] env[61936]: DEBUG nova.compute.manager [req-1800eff8-e629-436b-830e-63de334d3ecd req-581b5d6f-8fec-496d-a239-92c4d2d29f75 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] No waiting events found dispatching network-vif-plugged-40113568-3584-481c-aeb0-65d87745501c {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1061.716563] env[61936]: WARNING nova.compute.manager [req-1800eff8-e629-436b-830e-63de334d3ecd req-581b5d6f-8fec-496d-a239-92c4d2d29f75 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Received unexpected event network-vif-plugged-40113568-3584-481c-aeb0-65d87745501c for instance with vm_state building and task_state spawning. [ 1061.836932] env[61936]: INFO nova.compute.manager [-] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Took 1.32 seconds to deallocate network for instance. [ 1061.839558] env[61936]: DEBUG nova.network.neutron [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1061.854901] env[61936]: DEBUG nova.network.neutron [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Successfully updated port: 40113568-3584-481c-aeb0-65d87745501c {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.979886] env[61936]: DEBUG nova.network.neutron [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Updating instance_info_cache with network_info: [{"id": "24d5309e-de7c-4a1c-a314-cb7858990b81", "address": "fa:16:3e:e1:68:fd", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24d5309e-de", "ovs_interfaceid": "24d5309e-de7c-4a1c-a314-cb7858990b81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.093848] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.156086] env[61936]: DEBUG nova.compute.utils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.158680] env[61936]: DEBUG nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1062.158680] env[61936]: DEBUG nova.network.neutron [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1062.171952] env[61936]: DEBUG nova.compute.manager [req-769f368b-876b-47d9-b629-9184645910b7 req-56d36bba-705c-45bf-ac77-1a3830104347 service nova] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Received event network-vif-deleted-21257d10-73cd-4b46-bc5d-f6f78281e576 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1062.199353] env[61936]: DEBUG nova.policy [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1062.349028] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.349028] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.349028] env[61936]: DEBUG nova.objects.instance [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lazy-loading 'resources' on Instance uuid 160d7915-3795-460d-961f-7b43c8ed0168 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.359193] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.359522] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.359766] env[61936]: DEBUG nova.network.neutron [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.479786] env[61936]: DEBUG nova.network.neutron [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Successfully created port: 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1062.483272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "refresh_cache-c6043bea-03be-427d-9f39-43f81b0788ea" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.483579] env[61936]: DEBUG nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Instance network_info: |[{"id": "24d5309e-de7c-4a1c-a314-cb7858990b81", "address": "fa:16:3e:e1:68:fd", "network": {"id": "cfe3d93d-2207-47c9-895a-33c25ef01ebe", "bridge": "br-int", "label": "tempest-ServersTestJSON-1691238494-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90c29449db6c4beaa7f190d9225cb08a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24d5309e-de", "ovs_interfaceid": "24d5309e-de7c-4a1c-a314-cb7858990b81", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1062.484021] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:68:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24d5309e-de7c-4a1c-a314-cb7858990b81', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.491856] env[61936]: DEBUG oslo.service.loopingcall [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.492138] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1062.492398] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c17f109-3cec-4bd1-b244-cd499d17f7b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.513357] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.513357] env[61936]: value = "task-1253476" [ 1062.513357] env[61936]: _type = "Task" [ 1062.513357] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.520939] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253476, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.593890] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253475, 'name': ReconfigVM_Task, 'duration_secs': 0.777164} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.594305] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Reconfigured VM instance instance-00000065 to attach disk [datastore1] bbb98011-2321-4a69-a882-a4d54df132cc/bbb98011-2321-4a69-a882-a4d54df132cc.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.594983] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ac937e7-0aae-4ee1-a704-e00a946e7746 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.602063] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1062.602063] env[61936]: value = "task-1253477" [ 1062.602063] env[61936]: _type = "Task" [ 1062.602063] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.610549] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253477, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.660250] env[61936]: DEBUG nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1062.906265] env[61936]: DEBUG nova.network.neutron [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1063.019299] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6f292e-e9d6-4aad-bbbf-5e8cc94e73c5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.031373] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c924b896-2230-438a-b012-d6206fadac67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.035882] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253476, 'name': CreateVM_Task, 'duration_secs': 0.352263} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.036523] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1063.037051] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.037232] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.037557] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1063.037809] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-156b02be-a489-43fa-85ef-610dfd586d39 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.064746] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a4a72c-0518-4679-bb8f-b4f4bba75412 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.068868] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1063.068868] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52faaa33-fcab-d8c3-06b2-ea795e55706b" [ 1063.068868] env[61936]: _type = "Task" [ 1063.068868] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.074839] env[61936]: DEBUG nova.network.neutron [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Updating instance_info_cache with network_info: [{"id": "40113568-3584-481c-aeb0-65d87745501c", "address": "fa:16:3e:4b:b7:5d", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40113568-35", "ovs_interfaceid": "40113568-3584-481c-aeb0-65d87745501c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.077345] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b83037d-6abd-4121-895e-45bb5890c156 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.084509] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52faaa33-fcab-d8c3-06b2-ea795e55706b, 'name': SearchDatastore_Task, 'duration_secs': 0.009342} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.085128] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.085370] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.085606] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.085756] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.085934] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.086193] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a9829fb-8781-417a-b94b-d88de3a1f772 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.095283] env[61936]: DEBUG nova.compute.provider_tree [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.102871] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1063.102871] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1063.106632] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b4b531b-cdc7-41a3-8580-bdbb3bfc8a5d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.114050] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253477, 'name': Rename_Task, 'duration_secs': 0.204554} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.115154] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.115429] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1063.115429] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5215f2d5-37a2-6ede-648a-ab0ff3b31518" [ 1063.115429] env[61936]: _type = "Task" [ 1063.115429] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.115721] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-915e8857-1bd2-4da2-8e6f-0cf2b79cc46b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.125899] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5215f2d5-37a2-6ede-648a-ab0ff3b31518, 'name': SearchDatastore_Task, 'duration_secs': 0.008124} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.127433] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1063.127433] env[61936]: value = "task-1253478" [ 1063.127433] env[61936]: _type = "Task" [ 1063.127433] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.127626] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-395e10c0-cb24-43ec-91e2-3068d0013b7a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.135116] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1063.135116] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52fb8e11-93ea-bfae-cd9e-7a051e0f1c0e" [ 1063.135116] env[61936]: _type = "Task" [ 1063.135116] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.137969] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253478, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.144944] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52fb8e11-93ea-bfae-cd9e-7a051e0f1c0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.211655] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1063.211897] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270035', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'name': 'volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b', 'attached_at': '', 'detached_at': '', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'serial': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1063.212818] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7be102d-5515-420d-b365-0b9a82737802 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.232031] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66461e65-5f1c-4fe9-b5ed-2038002406af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.256661] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779/volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.256914] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-814b4e69-aed1-4fdb-9ce3-1ecfb4cfffa8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.274299] env[61936]: DEBUG oslo_vmware.api [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1063.274299] env[61936]: value = "task-1253479" [ 1063.274299] env[61936]: _type = "Task" [ 1063.274299] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.283062] env[61936]: DEBUG oslo_vmware.api [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253479, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.585305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.585714] env[61936]: DEBUG nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Instance network_info: |[{"id": "40113568-3584-481c-aeb0-65d87745501c", "address": "fa:16:3e:4b:b7:5d", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40113568-35", "ovs_interfaceid": "40113568-3584-481c-aeb0-65d87745501c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1063.586308] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:b7:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40113568-3584-481c-aeb0-65d87745501c', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.594297] env[61936]: DEBUG oslo.service.loopingcall [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.594551] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.594788] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-891c1e7c-e11e-44bf-a462-917b0a4aa94d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.610155] env[61936]: DEBUG nova.scheduler.client.report [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1063.619303] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.619303] env[61936]: value = "task-1253480" [ 1063.619303] env[61936]: _type = "Task" [ 1063.619303] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.627254] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253480, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.641412] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253478, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.649597] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52fb8e11-93ea-bfae-cd9e-7a051e0f1c0e, 'name': SearchDatastore_Task, 'duration_secs': 0.008779} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.649851] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.650131] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] c6043bea-03be-427d-9f39-43f81b0788ea/c6043bea-03be-427d-9f39-43f81b0788ea.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1063.650415] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0bae88ca-57cc-459a-9224-89b352d25f7c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.656825] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1063.656825] env[61936]: value = "task-1253481" [ 1063.656825] env[61936]: _type = "Task" [ 1063.656825] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.664091] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.669723] env[61936]: DEBUG nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1063.697922] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1063.698283] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.698640] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1063.698752] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.698838] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1063.699012] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1063.699260] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1063.699478] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1063.699664] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1063.699844] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1063.700035] env[61936]: DEBUG nova.virt.hardware [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1063.701625] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ebb914-7159-4045-963c-87c2b76fa78f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.710405] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46e9cc3-a063-45a9-95bd-48327d6af29e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.744524] env[61936]: DEBUG nova.compute.manager [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Received event network-changed-40113568-3584-481c-aeb0-65d87745501c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1063.744727] env[61936]: DEBUG nova.compute.manager [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Refreshing instance network info cache due to event network-changed-40113568-3584-481c-aeb0-65d87745501c. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1063.744950] env[61936]: DEBUG oslo_concurrency.lockutils [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] Acquiring lock "refresh_cache-9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.745114] env[61936]: DEBUG oslo_concurrency.lockutils [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] Acquired lock "refresh_cache-9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.745278] env[61936]: DEBUG nova.network.neutron [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Refreshing network info cache for port 40113568-3584-481c-aeb0-65d87745501c {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1063.784648] env[61936]: DEBUG oslo_vmware.api [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253479, 'name': ReconfigVM_Task, 'duration_secs': 0.365708} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.784962] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Reconfigured VM instance instance-00000060 to attach disk [datastore1] volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779/volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.791135] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d8594c2-6fba-4249-8a20-36de7fc4eafc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.806678] env[61936]: DEBUG oslo_vmware.api [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1063.806678] env[61936]: value = "task-1253482" [ 1063.806678] env[61936]: _type = "Task" [ 1063.806678] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.814701] env[61936]: DEBUG oslo_vmware.api [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253482, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.001854] env[61936]: DEBUG nova.network.neutron [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Successfully updated port: 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.116300] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.134554] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253480, 'name': CreateVM_Task, 'duration_secs': 0.385078} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.141113] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.142272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.142468] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.143138] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1064.144682] env[61936]: INFO nova.scheduler.client.report [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Deleted allocations for instance 160d7915-3795-460d-961f-7b43c8ed0168 [ 1064.147252] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eedb797f-6cba-4545-9433-40a9f08ef75d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.161796] env[61936]: DEBUG oslo_vmware.api [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253478, 'name': PowerOnVM_Task, 'duration_secs': 0.600971} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.166834] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1064.167142] env[61936]: INFO nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Took 9.28 seconds to spawn the instance on the hypervisor. [ 1064.167337] env[61936]: DEBUG nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1064.169604] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2871d8-6bea-428f-a83c-d95cb63c763b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.172538] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1064.172538] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52091c72-db0b-ef33-ef39-ce95dbbb9a5a" [ 1064.172538] env[61936]: _type = "Task" [ 1064.172538] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.183685] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253481, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.189377] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52091c72-db0b-ef33-ef39-ce95dbbb9a5a, 'name': SearchDatastore_Task, 'duration_secs': 0.013895} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.189937] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.190046] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.190347] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.190639] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.190772] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.191065] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53f34df3-aa41-4f00-8cef-0d44403ec86b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.202162] env[61936]: DEBUG nova.compute.manager [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-vif-plugged-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1064.202446] env[61936]: DEBUG oslo_concurrency.lockutils [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] Acquiring lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.202665] env[61936]: DEBUG oslo_concurrency.lockutils [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.202880] env[61936]: DEBUG oslo_concurrency.lockutils [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.203101] env[61936]: DEBUG nova.compute.manager [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] No waiting events found dispatching network-vif-plugged-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1064.203321] env[61936]: WARNING nova.compute.manager [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received unexpected event network-vif-plugged-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d for instance with vm_state building and task_state spawning. [ 1064.203614] env[61936]: DEBUG nova.compute.manager [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1064.203887] env[61936]: DEBUG nova.compute.manager [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing instance network info cache due to event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1064.204092] env[61936]: DEBUG oslo_concurrency.lockutils [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.204241] env[61936]: DEBUG oslo_concurrency.lockutils [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.204403] env[61936]: DEBUG nova.network.neutron [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.207129] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.207317] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.208370] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b47fa2d7-d48b-433e-a0c4-ad4302e007cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.218287] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1064.218287] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5215e93d-0383-5029-faa2-956e12870b3e" [ 1064.218287] env[61936]: _type = "Task" [ 1064.218287] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.227784] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5215e93d-0383-5029-faa2-956e12870b3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.318869] env[61936]: DEBUG oslo_vmware.api [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253482, 'name': ReconfigVM_Task, 'duration_secs': 0.16352} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.319183] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270035', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'name': 'volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b', 'attached_at': '', 'detached_at': '', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'serial': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1064.440328] env[61936]: DEBUG nova.network.neutron [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Updated VIF entry in instance network info cache for port 40113568-3584-481c-aeb0-65d87745501c. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1064.440773] env[61936]: DEBUG nova.network.neutron [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Updating instance_info_cache with network_info: [{"id": "40113568-3584-481c-aeb0-65d87745501c", "address": "fa:16:3e:4b:b7:5d", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40113568-35", "ovs_interfaceid": "40113568-3584-481c-aeb0-65d87745501c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.504707] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.667671] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b8094c1f-6132-4f8b-8149-9f53c00b7513 tempest-ServerDiskConfigTestJSON-425733311 tempest-ServerDiskConfigTestJSON-425733311-project-member] Lock "160d7915-3795-460d-961f-7b43c8ed0168" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.831s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.672514] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253481, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541494} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.672816] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] c6043bea-03be-427d-9f39-43f81b0788ea/c6043bea-03be-427d-9f39-43f81b0788ea.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1064.673104] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.673408] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee5c64b9-c157-40c9-bd06-0bac84e83ef5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.679497] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1064.679497] env[61936]: value = "task-1253483" [ 1064.679497] env[61936]: _type = "Task" [ 1064.679497] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.691057] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.693759] env[61936]: INFO nova.compute.manager [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Took 18.76 seconds to build instance. [ 1064.729015] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5215e93d-0383-5029-faa2-956e12870b3e, 'name': SearchDatastore_Task, 'duration_secs': 0.018014} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.729786] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1fbbde2-9326-4112-bc4a-e97b0d5a0e80 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.735504] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1064.735504] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52eec7bd-378d-8c8f-2e11-c213d19e7d0a" [ 1064.735504] env[61936]: _type = "Task" [ 1064.735504] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.743873] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52eec7bd-378d-8c8f-2e11-c213d19e7d0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.755191] env[61936]: DEBUG nova.network.neutron [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1064.833842] env[61936]: DEBUG nova.network.neutron [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.943943] env[61936]: DEBUG oslo_concurrency.lockutils [req-2a67e31c-6209-46b3-af3c-a929faae8c4d req-69334da8-7369-482d-89fe-a4eafce8f908 service nova] Releasing lock "refresh_cache-9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.189030] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065234} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.189352] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.190096] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c725101-c6aa-4ae2-861f-2f60d51afe84 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.203601] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e3239260-1167-4b89-b3ba-ed3682bd87eb tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "bbb98011-2321-4a69-a882-a4d54df132cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.325s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.212568] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] c6043bea-03be-427d-9f39-43f81b0788ea/c6043bea-03be-427d-9f39-43f81b0788ea.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.213169] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce70006a-21de-4f97-8107-ee85aaa10ba3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.233707] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1065.233707] env[61936]: value = "task-1253484" [ 1065.233707] env[61936]: _type = "Task" [ 1065.233707] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.245852] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253484, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.249342] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52eec7bd-378d-8c8f-2e11-c213d19e7d0a, 'name': SearchDatastore_Task, 'duration_secs': 0.010549} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.249587] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.249843] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.250116] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fdce01d1-8392-4c0f-80cf-216bd720d1e7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.257070] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1065.257070] env[61936]: value = "task-1253485" [ 1065.257070] env[61936]: _type = "Task" [ 1065.257070] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.266088] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253485, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.336806] env[61936]: DEBUG oslo_concurrency.lockutils [req-2b6c4e13-aedc-4df5-8edb-f10473b3adcb req-b31d37fb-18a2-4fe3-b320-cadc452477e7 service nova] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.337224] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.337388] env[61936]: DEBUG nova.network.neutron [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.353533] env[61936]: DEBUG nova.objects.instance [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'flavor' on Instance uuid 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.666435] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041a94ad-185c-49b5-a1c4-3043c5a4cea3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.674553] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd2e774-544d-4a9d-9d56-9802d251dad4 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Suspending the VM {{(pid=61936) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1065.674553] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6734d9a3-b68e-4b99-b20f-c46321b170e9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.680997] env[61936]: DEBUG oslo_vmware.api [None req-bdd2e774-544d-4a9d-9d56-9802d251dad4 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1065.680997] env[61936]: value = "task-1253486" [ 1065.680997] env[61936]: _type = "Task" [ 1065.680997] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.690894] env[61936]: DEBUG oslo_vmware.api [None req-bdd2e774-544d-4a9d-9d56-9802d251dad4 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253486, 'name': SuspendVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.744051] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253484, 'name': ReconfigVM_Task, 'duration_secs': 0.443651} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.744407] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Reconfigured VM instance instance-00000066 to attach disk [datastore1] c6043bea-03be-427d-9f39-43f81b0788ea/c6043bea-03be-427d-9f39-43f81b0788ea.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.745115] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77bd1eed-8fcc-4bdb-9eb6-f62a0ff91eae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.751258] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1065.751258] env[61936]: value = "task-1253487" [ 1065.751258] env[61936]: _type = "Task" [ 1065.751258] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.759788] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253487, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.768117] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253485, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461055} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.768400] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1065.768644] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1065.768876] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cdbf5c9e-dd1a-4327-9507-e703de8ee10a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.775260] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1065.775260] env[61936]: value = "task-1253488" [ 1065.775260] env[61936]: _type = "Task" [ 1065.775260] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.783971] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253488, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.859432] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4d4dc3fb-bd3d-4414-8024-211f3e7a617d tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.275s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.873638] env[61936]: DEBUG nova.network.neutron [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1066.030116] env[61936]: DEBUG nova.network.neutron [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.133860] env[61936]: DEBUG oslo_concurrency.lockutils [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.134775] env[61936]: DEBUG oslo_concurrency.lockutils [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.190910] env[61936]: DEBUG oslo_vmware.api [None req-bdd2e774-544d-4a9d-9d56-9802d251dad4 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253486, 'name': SuspendVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.260855] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253487, 'name': Rename_Task, 'duration_secs': 0.358792} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.261198] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1066.261498] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a4ee824-2f45-451e-8580-536eada91d45 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.268277] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1066.268277] env[61936]: value = "task-1253489" [ 1066.268277] env[61936]: _type = "Task" [ 1066.268277] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.282634] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.287374] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253488, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123251} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.287594] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.288427] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b61373-6c0e-4349-90ea-0be0cd754e9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.291020] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.291266] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.312939] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.314245] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cb7fb39-32cc-48e7-bf5a-b6bea6936079 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.334802] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1066.334802] env[61936]: value = "task-1253490" [ 1066.334802] env[61936]: _type = "Task" [ 1066.334802] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.343549] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253490, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.532834] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.533202] env[61936]: DEBUG nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Instance network_info: |[{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1066.533654] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:b9:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.541359] env[61936]: DEBUG oslo.service.loopingcall [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.541606] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1066.541887] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef5389d4-be0d-47c4-a435-863fc424825f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.562057] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.562057] env[61936]: value = "task-1253491" [ 1066.562057] env[61936]: _type = "Task" [ 1066.562057] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.570088] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253491, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.638144] env[61936]: INFO nova.compute.manager [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Detaching volume 0d436f9e-4fc6-4a67-b16e-9d113e7ae779 [ 1066.677936] env[61936]: INFO nova.virt.block_device [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Attempting to driver detach volume 0d436f9e-4fc6-4a67-b16e-9d113e7ae779 from mountpoint /dev/sdb [ 1066.678268] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1066.678569] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270035', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'name': 'volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b', 'attached_at': '', 'detached_at': '', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'serial': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1066.679514] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0635f11-7efe-4322-be28-601597350770 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.707098] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cd2059-3924-4f9b-9177-2893120dee44 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.709869] env[61936]: DEBUG oslo_vmware.api [None req-bdd2e774-544d-4a9d-9d56-9802d251dad4 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253486, 'name': SuspendVM_Task, 'duration_secs': 0.640045} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.710184] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bdd2e774-544d-4a9d-9d56-9802d251dad4 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Suspended the VM {{(pid=61936) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1066.710570] env[61936]: DEBUG nova.compute.manager [None req-bdd2e774-544d-4a9d-9d56-9802d251dad4 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1066.711690] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f85b072-53b8-4bf2-bd44-47141c4e6ff7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.717253] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88889778-ff88-4a4b-9e86-57bc8581af79 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.744350] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a4c37b-d707-458b-ac48-ebfddd956a3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.761274] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] The volume has not been displaced from its original location: [datastore1] volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779/volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1066.766618] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1066.766983] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21473668-b35b-4dd0-b83c-7132b54b1c71 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.790782] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253489, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.792493] env[61936]: DEBUG oslo_vmware.api [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1066.792493] env[61936]: value = "task-1253492" [ 1066.792493] env[61936]: _type = "Task" [ 1066.792493] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.796158] env[61936]: INFO nova.compute.manager [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Detaching volume 1d559b9c-feac-4939-85bd-15d6203183b4 [ 1066.804180] env[61936]: DEBUG oslo_vmware.api [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253492, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.832216] env[61936]: INFO nova.virt.block_device [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Attempting to driver detach volume 1d559b9c-feac-4939-85bd-15d6203183b4 from mountpoint /dev/sdb [ 1066.832493] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1066.832714] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270021', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'name': 'volume-1d559b9c-feac-4939-85bd-15d6203183b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8', 'attached_at': '', 'detached_at': '', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'serial': '1d559b9c-feac-4939-85bd-15d6203183b4'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1066.833659] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e49042-8470-4c16-90a2-02b317679f4d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.861166] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253490, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.861997] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55e0912-7143-4a65-8cc3-d626f53b002b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.868951] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb95ed2-6471-42e1-bc47-adc4a774bb46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.889942] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf2850a-53b1-4741-ad61-77cfaa039ad9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.907482] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] The volume has not been displaced from its original location: [datastore1] volume-1d559b9c-feac-4939-85bd-15d6203183b4/volume-1d559b9c-feac-4939-85bd-15d6203183b4.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1066.912852] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1066.913226] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6245e4c9-5d8d-45e0-822d-e4cf63f06a02 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.930879] env[61936]: DEBUG oslo_vmware.api [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1066.930879] env[61936]: value = "task-1253493" [ 1066.930879] env[61936]: _type = "Task" [ 1066.930879] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.939568] env[61936]: DEBUG oslo_vmware.api [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253493, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.073018] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253491, 'name': CreateVM_Task, 'duration_secs': 0.376452} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.073018] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.073428] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.073642] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.073997] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.074304] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9638f582-eb35-442c-916c-e158a6befed5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.079112] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1067.079112] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e03ef6-9330-0f9a-58ab-a876530f1ccf" [ 1067.079112] env[61936]: _type = "Task" [ 1067.079112] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.087753] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e03ef6-9330-0f9a-58ab-a876530f1ccf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.291423] env[61936]: DEBUG oslo_vmware.api [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253489, 'name': PowerOnVM_Task, 'duration_secs': 0.934311} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.291858] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.292055] env[61936]: INFO nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Took 8.20 seconds to spawn the instance on the hypervisor. [ 1067.292243] env[61936]: DEBUG nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1067.293086] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f764d000-5fa9-4a33-b7a6-acea16f0fdc7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.309011] env[61936]: DEBUG oslo_vmware.api [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253492, 'name': ReconfigVM_Task, 'duration_secs': 0.244693} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.309636] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1067.314982] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a89f9bf0-f99f-4b66-8e5e-5e60cf4c067e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.334480] env[61936]: DEBUG oslo_vmware.api [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1067.334480] env[61936]: value = "task-1253494" [ 1067.334480] env[61936]: _type = "Task" [ 1067.334480] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.346880] env[61936]: DEBUG oslo_vmware.api [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253494, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.350411] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253490, 'name': ReconfigVM_Task, 'duration_secs': 0.812801} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.350748] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.351496] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4e3d624-148c-47e2-ae93-73fbcec3d205 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.359343] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1067.359343] env[61936]: value = "task-1253495" [ 1067.359343] env[61936]: _type = "Task" [ 1067.359343] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.367891] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253495, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.441080] env[61936]: DEBUG oslo_vmware.api [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253493, 'name': ReconfigVM_Task, 'duration_secs': 0.215247} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.441441] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1067.446046] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97e4df11-1422-4d32-8e81-593c98f0dedd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.461276] env[61936]: DEBUG oslo_vmware.api [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1067.461276] env[61936]: value = "task-1253496" [ 1067.461276] env[61936]: _type = "Task" [ 1067.461276] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.470514] env[61936]: DEBUG oslo_vmware.api [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253496, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.593365] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e03ef6-9330-0f9a-58ab-a876530f1ccf, 'name': SearchDatastore_Task, 'duration_secs': 0.009859} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.593709] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.593948] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.594254] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.594483] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.594718] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.595284] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ff1dd0f-01ce-41b3-b0bd-2a97e4abe480 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.604517] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.604721] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.605719] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccd71f27-e432-4620-9314-a294ac655405 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.611183] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1067.611183] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e98adc-bc5d-2020-ce84-da1546325ec7" [ 1067.611183] env[61936]: _type = "Task" [ 1067.611183] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.621243] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e98adc-bc5d-2020-ce84-da1546325ec7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.816804] env[61936]: INFO nova.compute.manager [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Took 13.95 seconds to build instance. [ 1067.847995] env[61936]: DEBUG oslo_vmware.api [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253494, 'name': ReconfigVM_Task, 'duration_secs': 0.17182} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.848338] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270035', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'name': 'volume-0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b', 'attached_at': '', 'detached_at': '', 'volume_id': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779', 'serial': '0d436f9e-4fc6-4a67-b16e-9d113e7ae779'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1067.869531] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253495, 'name': Rename_Task, 'duration_secs': 0.20694} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.869858] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.870137] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0acfc4b2-a84c-4a4f-8ea9-d2f3eec525bb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.878465] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1067.878465] env[61936]: value = "task-1253497" [ 1067.878465] env[61936]: _type = "Task" [ 1067.878465] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.886951] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.970480] env[61936]: DEBUG oslo_vmware.api [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253496, 'name': ReconfigVM_Task, 'duration_secs': 0.148292} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.970818] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270021', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'name': 'volume-1d559b9c-feac-4939-85bd-15d6203183b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8', 'attached_at': '', 'detached_at': '', 'volume_id': '1d559b9c-feac-4939-85bd-15d6203183b4', 'serial': '1d559b9c-feac-4939-85bd-15d6203183b4'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1068.122351] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e98adc-bc5d-2020-ce84-da1546325ec7, 'name': SearchDatastore_Task, 'duration_secs': 0.008695} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.123089] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8344aac6-0d3c-4109-ada2-60093e4c4a27 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.128385] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1068.128385] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5251cec3-1bcd-bc35-f538-5d22efc1e334" [ 1068.128385] env[61936]: _type = "Task" [ 1068.128385] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.136913] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5251cec3-1bcd-bc35-f538-5d22efc1e334, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.321162] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e6316c4b-b439-443b-a794-91596e93aceb tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.464s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.388109] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253497, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.390861] env[61936]: DEBUG nova.objects.instance [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'flavor' on Instance uuid 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.505624] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "bbb98011-2321-4a69-a882-a4d54df132cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.505981] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "bbb98011-2321-4a69-a882-a4d54df132cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.506252] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "bbb98011-2321-4a69-a882-a4d54df132cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.506476] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "bbb98011-2321-4a69-a882-a4d54df132cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.506656] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "bbb98011-2321-4a69-a882-a4d54df132cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.509042] env[61936]: INFO nova.compute.manager [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Terminating instance [ 1068.516043] env[61936]: DEBUG nova.objects.instance [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.569305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "c6043bea-03be-427d-9f39-43f81b0788ea" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.569603] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.569793] env[61936]: DEBUG nova.compute.manager [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1068.570766] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c245456-b091-4ac4-8847-77393cbd4d89 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.577844] env[61936]: DEBUG nova.compute.manager [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61936) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1068.578411] env[61936]: DEBUG nova.objects.instance [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lazy-loading 'flavor' on Instance uuid c6043bea-03be-427d-9f39-43f81b0788ea {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.643332] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5251cec3-1bcd-bc35-f538-5d22efc1e334, 'name': SearchDatastore_Task, 'duration_secs': 0.008924} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.643759] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.644174] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f1d45dcb-768d-447e-be49-ef43ad98fa61/f1d45dcb-768d-447e-be49-ef43ad98fa61.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.644549] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-655b212e-fb46-49c6-a766-664526def72a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.652259] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1068.652259] env[61936]: value = "task-1253498" [ 1068.652259] env[61936]: _type = "Task" [ 1068.652259] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.665228] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.890270] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253497, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.012895] env[61936]: DEBUG nova.compute.manager [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1069.013168] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1069.014074] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8d3837-0b1e-46b8-9d6f-3b72fff10218 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.025753] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.026330] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23db4c42-4781-4cae-b568-43cf067891e4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.164681] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253498, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.197195] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.197195] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.197507] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleting the datastore file [datastore1] bbb98011-2321-4a69-a882-a4d54df132cc {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.198036] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf403a71-aa94-45af-b2fe-9eb1ee68bd18 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.206057] env[61936]: DEBUG oslo_vmware.api [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1069.206057] env[61936]: value = "task-1253500" [ 1069.206057] env[61936]: _type = "Task" [ 1069.206057] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.215232] env[61936]: DEBUG oslo_vmware.api [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253500, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.389660] env[61936]: DEBUG oslo_vmware.api [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253497, 'name': PowerOnVM_Task, 'duration_secs': 1.416568} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.389989] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1069.390103] env[61936]: INFO nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1069.390292] env[61936]: DEBUG nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1069.391118] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abd1491-3227-4fa7-a326-7c97110555b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.399869] env[61936]: DEBUG oslo_concurrency.lockutils [None req-544e5ab5-6b81-4978-8e15-d0f6d9655fef tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.265s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.496366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.524209] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ca2bc295-b2b4-4fbe-b24a-db2831c8647e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.233s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.525361] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.029s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.525604] env[61936]: DEBUG nova.compute.manager [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1069.526836] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0540a4-0439-42e6-976c-30380c26eee6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.534540] env[61936]: DEBUG nova.compute.manager [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61936) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1069.535136] env[61936]: DEBUG nova.objects.instance [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.586710] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1069.587413] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9fb1000-d6be-4d64-95b3-1c5090ac72d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.594645] env[61936]: DEBUG oslo_vmware.api [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1069.594645] env[61936]: value = "task-1253501" [ 1069.594645] env[61936]: _type = "Task" [ 1069.594645] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.602718] env[61936]: DEBUG oslo_vmware.api [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.663977] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574596} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.664230] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] f1d45dcb-768d-447e-be49-ef43ad98fa61/f1d45dcb-768d-447e-be49-ef43ad98fa61.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.664422] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1069.664673] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14d44d56-5aff-4213-82c3-e2975b9853cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.671404] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1069.671404] env[61936]: value = "task-1253502" [ 1069.671404] env[61936]: _type = "Task" [ 1069.671404] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.681215] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.715100] env[61936]: DEBUG oslo_vmware.api [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253500, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162255} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.715389] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.715577] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.715756] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.715935] env[61936]: INFO nova.compute.manager [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Took 0.70 seconds to destroy the instance on the hypervisor. [ 1069.716197] env[61936]: DEBUG oslo.service.loopingcall [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.716401] env[61936]: DEBUG nova.compute.manager [-] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1069.716497] env[61936]: DEBUG nova.network.neutron [-] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.910097] env[61936]: INFO nova.compute.manager [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Took 13.16 seconds to build instance. [ 1070.038315] env[61936]: DEBUG nova.compute.manager [req-6da1160e-e9e2-4127-855d-66b0e04b5dc8 req-ccb4a88c-7789-4adf-90d2-5aa2e422f181 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Received event network-vif-deleted-988c6387-6b32-4fc0-a731-70a6a97bc815 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1070.038637] env[61936]: INFO nova.compute.manager [req-6da1160e-e9e2-4127-855d-66b0e04b5dc8 req-ccb4a88c-7789-4adf-90d2-5aa2e422f181 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Neutron deleted interface 988c6387-6b32-4fc0-a731-70a6a97bc815; detaching it from the instance and deleting it from the info cache [ 1070.038893] env[61936]: DEBUG nova.network.neutron [req-6da1160e-e9e2-4127-855d-66b0e04b5dc8 req-ccb4a88c-7789-4adf-90d2-5aa2e422f181 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.104303] env[61936]: DEBUG oslo_vmware.api [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253501, 'name': PowerOffVM_Task, 'duration_secs': 0.461087} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.104576] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1070.104779] env[61936]: DEBUG nova.compute.manager [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1070.105555] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ddd9e2-9a39-4c36-b601-8fc957aaadf0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.180265] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064281} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.180525] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1070.181329] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a8db2d-ec0f-4690-a6ea-e295fd10f3ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.202598] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] f1d45dcb-768d-447e-be49-ef43ad98fa61/f1d45dcb-768d-447e-be49-ef43ad98fa61.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.202863] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8587848-b752-4034-b7e1-9cebbcf715c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.222553] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1070.222553] env[61936]: value = "task-1253503" [ 1070.222553] env[61936]: _type = "Task" [ 1070.222553] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.230341] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253503, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.412088] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f73e8b76-30bb-41ff-86ee-dad2d04c84d0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.666s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.508961] env[61936]: DEBUG nova.network.neutron [-] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.543019] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a24757b-f95c-408a-9c02-b2e21daf347c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.546553] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.546799] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0cd83d1b-5740-4642-8159-9ab434fbf9ce {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.549893] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.550146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.550361] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.550659] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.550848] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.553946] env[61936]: INFO nova.compute.manager [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Terminating instance [ 1070.558089] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0595266f-630d-4306-aa66-4d0a5518f5be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.570060] env[61936]: DEBUG oslo_vmware.api [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1070.570060] env[61936]: value = "task-1253504" [ 1070.570060] env[61936]: _type = "Task" [ 1070.570060] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.579696] env[61936]: DEBUG oslo_vmware.api [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253504, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.590887] env[61936]: DEBUG nova.compute.manager [req-6da1160e-e9e2-4127-855d-66b0e04b5dc8 req-ccb4a88c-7789-4adf-90d2-5aa2e422f181 service nova] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Detach interface failed, port_id=988c6387-6b32-4fc0-a731-70a6a97bc815, reason: Instance bbb98011-2321-4a69-a882-a4d54df132cc could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1070.617043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3380000d-e6f1-4b11-9ff8-b5d15ce5f09f tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.047s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.734481] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253503, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.012089] env[61936]: INFO nova.compute.manager [-] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Took 1.29 seconds to deallocate network for instance. [ 1071.076188] env[61936]: DEBUG nova.compute.manager [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1071.076414] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1071.077229] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b0bec5-b7db-4682-900a-b26586613240 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.084965] env[61936]: DEBUG oslo_vmware.api [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253504, 'name': PowerOffVM_Task, 'duration_secs': 0.338533} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.086913] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.087131] env[61936]: DEBUG nova.compute.manager [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1071.087411] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.088110] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d78717-c264-49a4-8a72-9a02811c2859 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.090313] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89580256-503d-4638-a60c-77004c8e7728 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.098029] env[61936]: DEBUG oslo_vmware.api [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1071.098029] env[61936]: value = "task-1253505" [ 1071.098029] env[61936]: _type = "Task" [ 1071.098029] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.104204] env[61936]: DEBUG oslo_vmware.api [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253505, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.159533] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.159804] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.159991] env[61936]: DEBUG nova.compute.manager [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1071.160976] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a3cb5f-93ec-4e6e-8134-3c97d7318911 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.167427] env[61936]: DEBUG nova.compute.manager [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61936) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1071.167974] env[61936]: DEBUG nova.objects.instance [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'flavor' on Instance uuid 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.234446] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253503, 'name': ReconfigVM_Task, 'duration_secs': 0.771592} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.234824] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Reconfigured VM instance instance-00000068 to attach disk [datastore1] f1d45dcb-768d-447e-be49-ef43ad98fa61/f1d45dcb-768d-447e-be49-ef43ad98fa61.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.235535] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e50aa710-87a0-4803-a1a1-1c6917d12229 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.241937] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1071.241937] env[61936]: value = "task-1253506" [ 1071.241937] env[61936]: _type = "Task" [ 1071.241937] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.250164] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253506, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.493843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "c6043bea-03be-427d-9f39-43f81b0788ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.494153] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.494366] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "c6043bea-03be-427d-9f39-43f81b0788ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.494550] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.494722] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.496850] env[61936]: INFO nova.compute.manager [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Terminating instance [ 1071.517775] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.517978] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.518249] env[61936]: DEBUG nova.objects.instance [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'resources' on Instance uuid bbb98011-2321-4a69-a882-a4d54df132cc {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.604732] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b68335cd-b0ee-454c-a56b-1ffe5a0d2092 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.079s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.610816] env[61936]: DEBUG oslo_vmware.api [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253505, 'name': PowerOffVM_Task, 'duration_secs': 0.140154} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.610816] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.610983] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1071.611178] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24c92201-971f-4f8f-ba25-297254bf2d9c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.674628] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1071.674864] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1071.675059] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleting the datastore file [datastore1] 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1071.675462] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2975511-9738-4898-baea-62cda49c85eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.681762] env[61936]: DEBUG oslo_vmware.api [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1071.681762] env[61936]: value = "task-1253508" [ 1071.681762] env[61936]: _type = "Task" [ 1071.681762] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.689887] env[61936]: DEBUG oslo_vmware.api [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253508, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.751588] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253506, 'name': Rename_Task, 'duration_secs': 0.130046} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.751953] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.752263] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f571bc2f-e997-4b5f-8404-6a0a33d0f611 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.758488] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1071.758488] env[61936]: value = "task-1253509" [ 1071.758488] env[61936]: _type = "Task" [ 1071.758488] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.767917] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.001410] env[61936]: DEBUG nova.compute.manager [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1072.001657] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1072.002636] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520d54ec-5819-4ba6-a332-e7319c016ee7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.010516] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1072.010763] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b41427ce-af98-4f10-84a5-4a27c1788726 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.045084] env[61936]: DEBUG nova.objects.instance [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.083232] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1072.083457] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1072.083643] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleting the datastore file [datastore1] c6043bea-03be-427d-9f39-43f81b0788ea {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.083903] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-472d5ae8-d941-4a4e-8324-92003a3dbb4d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.090692] env[61936]: DEBUG oslo_vmware.api [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for the task: (returnval){ [ 1072.090692] env[61936]: value = "task-1253511" [ 1072.090692] env[61936]: _type = "Task" [ 1072.090692] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.100769] env[61936]: DEBUG oslo_vmware.api [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253511, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.176190] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.177546] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ead1acdd-4da2-421c-9d25-69427dda4554 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.179884] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f86dd6-9619-4cce-afdd-bd57b1109ca2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.188541] env[61936]: DEBUG oslo_vmware.api [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1072.188541] env[61936]: value = "task-1253512" [ 1072.188541] env[61936]: _type = "Task" [ 1072.188541] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.194764] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359c0dfd-c335-4c2f-8559-e959081abb7b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.201860] env[61936]: DEBUG oslo_vmware.api [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253508, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131318} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.202857] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.203083] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1072.203262] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1072.203458] env[61936]: INFO nova.compute.manager [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1072.203715] env[61936]: DEBUG oslo.service.loopingcall [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.203920] env[61936]: DEBUG nova.compute.manager [-] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1072.204019] env[61936]: DEBUG nova.network.neutron [-] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1072.234915] env[61936]: DEBUG oslo_vmware.api [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253512, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.236382] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad28def-165d-4278-acc8-4a4d2a5ffa5f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.244714] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d793a600-aad3-4d5c-8443-0362ade6b33a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.259039] env[61936]: DEBUG nova.compute.provider_tree [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.268668] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253509, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.551604] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.552042] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.552160] env[61936]: DEBUG nova.network.neutron [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1072.552344] env[61936]: DEBUG nova.objects.instance [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'info_cache' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.602418] env[61936]: DEBUG oslo_vmware.api [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Task: {'id': task-1253511, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132642} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.603130] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1072.603242] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1072.603526] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1072.603616] env[61936]: INFO nova.compute.manager [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1072.604682] env[61936]: DEBUG oslo.service.loopingcall [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.604682] env[61936]: DEBUG nova.compute.manager [-] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1072.604682] env[61936]: DEBUG nova.network.neutron [-] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1072.700759] env[61936]: DEBUG oslo_vmware.api [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253512, 'name': PowerOffVM_Task, 'duration_secs': 0.200306} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.705482] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1072.705482] env[61936]: DEBUG nova.compute.manager [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1072.705482] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e74091-fef4-4fc2-b9ed-9486961a7d49 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.731937] env[61936]: DEBUG nova.compute.manager [req-3267a469-6463-4331-ad92-9c165ae611ce req-2cb3deab-eb93-4411-9095-bb9b75a01798 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Received event network-vif-deleted-b556882a-8fa4-4ea2-871b-6784ef87c88c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1072.733442] env[61936]: INFO nova.compute.manager [req-3267a469-6463-4331-ad92-9c165ae611ce req-2cb3deab-eb93-4411-9095-bb9b75a01798 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Neutron deleted interface b556882a-8fa4-4ea2-871b-6784ef87c88c; detaching it from the instance and deleting it from the info cache [ 1072.733442] env[61936]: DEBUG nova.network.neutron [req-3267a469-6463-4331-ad92-9c165ae611ce req-2cb3deab-eb93-4411-9095-bb9b75a01798 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.767702] env[61936]: DEBUG nova.scheduler.client.report [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1072.778243] env[61936]: DEBUG oslo_vmware.api [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253509, 'name': PowerOnVM_Task, 'duration_secs': 0.528343} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.778582] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1072.779313] env[61936]: INFO nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Took 9.11 seconds to spawn the instance on the hypervisor. [ 1072.779313] env[61936]: DEBUG nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1072.780139] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6810c4-0574-4cfe-9022-22b8a4cd939f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.907275] env[61936]: DEBUG nova.compute.manager [req-d298de8d-1d02-4682-be0a-83516fa4597e req-a5333327-86d8-4fc8-928b-654db2a81eef service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Received event network-vif-deleted-24d5309e-de7c-4a1c-a314-cb7858990b81 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1072.907494] env[61936]: INFO nova.compute.manager [req-d298de8d-1d02-4682-be0a-83516fa4597e req-a5333327-86d8-4fc8-928b-654db2a81eef service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Neutron deleted interface 24d5309e-de7c-4a1c-a314-cb7858990b81; detaching it from the instance and deleting it from the info cache [ 1072.907752] env[61936]: DEBUG nova.network.neutron [req-d298de8d-1d02-4682-be0a-83516fa4597e req-a5333327-86d8-4fc8-928b-654db2a81eef service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.056242] env[61936]: DEBUG nova.objects.base [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Object Instance<0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8> lazy-loaded attributes: flavor,info_cache {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1073.203716] env[61936]: DEBUG nova.network.neutron [-] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.213502] env[61936]: DEBUG oslo_concurrency.lockutils [None req-3e272c4b-40f6-4949-8dd0-a4f689552552 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.235985] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f380d07-d0a1-46b5-9c75-7093a8886855 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.247095] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d78e22-9dc8-4f9f-bf47-4d46fb75a685 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.276490] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.278619] env[61936]: DEBUG nova.compute.manager [req-3267a469-6463-4331-ad92-9c165ae611ce req-2cb3deab-eb93-4411-9095-bb9b75a01798 service nova] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Detach interface failed, port_id=b556882a-8fa4-4ea2-871b-6784ef87c88c, reason: Instance 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1073.299857] env[61936]: INFO nova.scheduler.client.report [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted allocations for instance bbb98011-2321-4a69-a882-a4d54df132cc [ 1073.301018] env[61936]: INFO nova.compute.manager [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Took 14.05 seconds to build instance. [ 1073.385083] env[61936]: DEBUG nova.network.neutron [-] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.410142] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b87f1e6-e552-42cf-b5d5-72d01834cb19 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.420243] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1dd3f9e-1c25-4971-87d3-4014ea26f6a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.452403] env[61936]: DEBUG nova.compute.manager [req-d298de8d-1d02-4682-be0a-83516fa4597e req-a5333327-86d8-4fc8-928b-654db2a81eef service nova] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Detach interface failed, port_id=24d5309e-de7c-4a1c-a314-cb7858990b81, reason: Instance c6043bea-03be-427d-9f39-43f81b0788ea could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1073.708525] env[61936]: INFO nova.compute.manager [-] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Took 1.50 seconds to deallocate network for instance. [ 1073.713950] env[61936]: INFO nova.compute.manager [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Rebuilding instance [ 1073.760211] env[61936]: DEBUG nova.compute.manager [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1073.760740] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e792896f-6bb3-416b-8f8d-a0bb5b6ad620 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.805674] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78179059-426d-495e-b8e0-e4fbea0a4c6a tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.566s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.808318] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e0fd7d99-2b01-4f53-a33a-197dc08a5126 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "bbb98011-2321-4a69-a882-a4d54df132cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.302s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.888051] env[61936]: INFO nova.compute.manager [-] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Took 1.28 seconds to deallocate network for instance. [ 1073.926391] env[61936]: DEBUG nova.network.neutron [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating instance_info_cache with network_info: [{"id": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "address": "fa:16:3e:ce:3d:0e", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap746d118b-a8", "ovs_interfaceid": "746d118b-a889-44a8-a6fe-1f8e1c483c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.217476] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.217800] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.218087] env[61936]: DEBUG nova.objects.instance [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'resources' on Instance uuid 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.394544] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.428871] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "refresh_cache-0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.754438] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.754850] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.761177] env[61936]: DEBUG nova.compute.manager [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1074.761377] env[61936]: DEBUG nova.compute.manager [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing instance network info cache due to event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1074.761723] env[61936]: DEBUG oslo_concurrency.lockutils [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.761931] env[61936]: DEBUG oslo_concurrency.lockutils [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.762128] env[61936]: DEBUG nova.network.neutron [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1074.779609] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1074.780715] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72037341-4a15-40bd-87f3-ca58fccf5d99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.788504] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1074.788504] env[61936]: value = "task-1253513" [ 1074.788504] env[61936]: _type = "Task" [ 1074.788504] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.802016] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1074.802297] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1074.803376] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ef13df-9aa7-44f7-8aee-38c8a4a5a829 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.810764] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.811046] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3822031-b66a-4853-b4fb-0905bd40ab55 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.877063] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.877458] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.877821] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleting the datastore file [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.878350] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b150303-4e25-4113-aa54-f1159bfe2008 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.888224] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1074.888224] env[61936]: value = "task-1253515" [ 1074.888224] env[61936]: _type = "Task" [ 1074.888224] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.898121] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253515, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.899735] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655b9975-03dd-47d5-ad1a-e9c26335b7b5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.908461] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e73e39-d829-4247-930f-6b3d78e16c07 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.943856] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346c07b9-c65f-4e8d-8f4f-7b466c1ed963 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.951252] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362cdb7b-ef1f-4d0b-899b-ad4fcc1a72b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.965251] env[61936]: DEBUG nova.compute.provider_tree [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.261464] env[61936]: DEBUG nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1075.362574] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.362814] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.398135] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253515, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135741} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.400545] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.400743] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.400924] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.443859] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.444469] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c145ee21-8d43-4c0a-96a1-523779756436 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.452130] env[61936]: DEBUG oslo_vmware.api [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1075.452130] env[61936]: value = "task-1253516" [ 1075.452130] env[61936]: _type = "Task" [ 1075.452130] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.460411] env[61936]: DEBUG oslo_vmware.api [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253516, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.468438] env[61936]: DEBUG nova.scheduler.client.report [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1075.533792] env[61936]: DEBUG nova.network.neutron [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updated VIF entry in instance network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1075.534267] env[61936]: DEBUG nova.network.neutron [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.784614] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.864891] env[61936]: DEBUG nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1075.962395] env[61936]: DEBUG oslo_vmware.api [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253516, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.975612] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.977615] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.583s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.977849] env[61936]: DEBUG nova.objects.instance [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lazy-loading 'resources' on Instance uuid c6043bea-03be-427d-9f39-43f81b0788ea {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.993404] env[61936]: INFO nova.scheduler.client.report [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted allocations for instance 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b [ 1076.036473] env[61936]: DEBUG oslo_concurrency.lockutils [req-5dde8073-d211-4722-b38b-23e79d5c2ea3 req-13059920-af21-4ae0-bfaa-f1fccef9af08 service nova] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.383047] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.432915] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1076.433242] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.433381] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1076.433567] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.433717] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1076.433864] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1076.434085] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1076.434255] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1076.434425] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1076.434591] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1076.434764] env[61936]: DEBUG nova.virt.hardware [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1076.435638] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad41792-e056-488d-83ab-f4988af89590 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.443750] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7768ef95-25b6-4160-b8aa-fedc63bb2f39 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.457382] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:b7:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40113568-3584-481c-aeb0-65d87745501c', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.464744] env[61936]: DEBUG oslo.service.loopingcall [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.467753] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.468009] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c097191f-9424-4ebb-9207-26cc4baafa18 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.489410] env[61936]: DEBUG oslo_vmware.api [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253516, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.493624] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.493624] env[61936]: value = "task-1253517" [ 1076.493624] env[61936]: _type = "Task" [ 1076.493624] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.504554] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253517, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.505021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8c85996a-33c2-4cc4-8338-268531fa840e tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.955s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.625623] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64663632-0cd2-4d16-bab2-dc27c5885714 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.633641] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb0655a-2d17-4f9c-bb6e-576ffbee5bd3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.664433] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e638e46-80d5-4f5b-9aca-1fd1f460f1b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.672392] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062e36c9-ca6b-48f5-aa80-b13227ea6d1f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.687903] env[61936]: DEBUG nova.compute.provider_tree [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.966749] env[61936]: DEBUG oslo_vmware.api [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253516, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.003461] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253517, 'name': CreateVM_Task, 'duration_secs': 0.329245} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.003566] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.004230] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.004422] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.004722] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1077.005923] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4083b7ed-cf22-4259-be0b-c674e52e23f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.011620] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1077.011620] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ab4c8f-8632-b395-8d52-dedb2fa32057" [ 1077.011620] env[61936]: _type = "Task" [ 1077.011620] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.018800] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ab4c8f-8632-b395-8d52-dedb2fa32057, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.191154] env[61936]: DEBUG nova.scheduler.client.report [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1077.468255] env[61936]: DEBUG oslo_vmware.api [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253516, 'name': PowerOnVM_Task, 'duration_secs': 1.908874} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.468526] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1077.468724] env[61936]: DEBUG nova.compute.manager [None req-b7160bba-dc4d-44c5-bd08-be99470f40eb tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1077.469474] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d6aa63-9101-4052-a022-371307c39c03 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.522595] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ab4c8f-8632-b395-8d52-dedb2fa32057, 'name': SearchDatastore_Task, 'duration_secs': 0.00887} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.522891] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.523142] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1077.523379] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.523536] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.523789] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.523968] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-750176c2-bc83-4d4c-bf7f-0376b0ac4319 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.532102] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.532294] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1077.533026] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56aba477-be7e-4951-ad20-a3fff2bfc274 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.538141] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1077.538141] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523cf77c-e932-ae13-b83e-ca4cc0490f1e" [ 1077.538141] env[61936]: _type = "Task" [ 1077.538141] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.545353] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523cf77c-e932-ae13-b83e-ca4cc0490f1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.697086] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.699105] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.915s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.700938] env[61936]: INFO nova.compute.claims [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.720847] env[61936]: INFO nova.scheduler.client.report [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Deleted allocations for instance c6043bea-03be-427d-9f39-43f81b0788ea [ 1078.049587] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523cf77c-e932-ae13-b83e-ca4cc0490f1e, 'name': SearchDatastore_Task, 'duration_secs': 0.007961} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.050437] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70ddd8b7-4396-4944-8e47-1dacc50ee043 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.055617] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1078.055617] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d95bf8-a3f4-cb14-ec2a-5fb1e01de0bc" [ 1078.055617] env[61936]: _type = "Task" [ 1078.055617] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.064339] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d95bf8-a3f4-cb14-ec2a-5fb1e01de0bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.229817] env[61936]: DEBUG oslo_concurrency.lockutils [None req-db8d2538-b546-40b8-94b8-3fdf537c5db3 tempest-ServersTestJSON-306269860 tempest-ServersTestJSON-306269860-project-member] Lock "c6043bea-03be-427d-9f39-43f81b0788ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.736s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.567151] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d95bf8-a3f4-cb14-ec2a-5fb1e01de0bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009326} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.567441] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.567723] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1078.567984] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43474653-4653-4f72-9bd5-fdad5f52ea9f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.574597] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1078.574597] env[61936]: value = "task-1253518" [ 1078.574597] env[61936]: _type = "Task" [ 1078.574597] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.582712] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253518, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.669679] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.669998] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.858824] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3397c79-ac64-4d38-8e14-dce05453249e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.867261] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73744806-c953-43ef-8b96-95fae29e8e77 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.900714] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbd2a07-e6ba-4801-ba1d-afed5cd17cb8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.908517] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddecef1-2cb8-4a7e-bb59-bfd4d963d86f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.924689] env[61936]: DEBUG nova.compute.provider_tree [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.085383] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253518, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.174896] env[61936]: DEBUG nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1079.428332] env[61936]: DEBUG nova.scheduler.client.report [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1079.585953] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253518, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590395} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.586231] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1079.586533] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1079.586814] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8cb609a-8a13-4a53-a334-881e721872a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.594074] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1079.594074] env[61936]: value = "task-1253519" [ 1079.594074] env[61936]: _type = "Task" [ 1079.594074] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.604928] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253519, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.696949] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.942063] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.942981] env[61936]: DEBUG nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1079.946131] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.563s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.947982] env[61936]: INFO nova.compute.claims [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1080.107514] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253519, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070469} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.107862] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1080.108688] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b895dd95-b52b-4b96-a9e0-8faa6702f249 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.140658] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.141405] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24d11f5b-e53a-4005-938f-62e4376fcd9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.162768] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1080.162768] env[61936]: value = "task-1253520" [ 1080.162768] env[61936]: _type = "Task" [ 1080.162768] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.172676] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253520, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.452793] env[61936]: DEBUG nova.compute.utils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1080.457641] env[61936]: DEBUG nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1080.457641] env[61936]: DEBUG nova.network.neutron [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1080.504140] env[61936]: DEBUG nova.policy [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcee2be71ae044338df4422a3c575ca4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e81b119251964bdb8e5a0ea84b29f2a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1080.676995] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.878229] env[61936]: DEBUG nova.network.neutron [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Successfully created port: c0d12433-0c29-4d8f-9664-dcbb76583dfc {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.958831] env[61936]: DEBUG nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1081.164081] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a499a4bd-acca-4ec9-9b66-b7ab3efa1446 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.178821] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd83512-3154-4be5-8fbb-fad6a088a7aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.184795] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253520, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.232728] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dcf89a-a067-4025-8059-c42965d6e448 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.246027] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3ffc09-5a42-40dc-a079-dab519ec888d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.267273] env[61936]: DEBUG nova.compute.provider_tree [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.678883] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253520, 'name': ReconfigVM_Task, 'duration_secs': 1.122837} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.678883] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc/9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.679409] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ecc9f5d-a1f6-4614-a5ec-ab6d7635b06a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.685984] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1081.685984] env[61936]: value = "task-1253521" [ 1081.685984] env[61936]: _type = "Task" [ 1081.685984] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.693449] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253521, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.771367] env[61936]: DEBUG nova.scheduler.client.report [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1081.974051] env[61936]: DEBUG nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1081.999694] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1082.000679] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.000679] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1082.000679] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.000679] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1082.000679] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1082.000916] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1082.001096] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1082.001310] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1082.001483] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1082.001669] env[61936]: DEBUG nova.virt.hardware [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1082.003026] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca17ac5-e842-4981-99a8-81e32d642100 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.011397] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67f21f6-7a66-4e9d-8b15-5fe6ba4855c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.195638] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253521, 'name': Rename_Task, 'duration_secs': 0.140973} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.195940] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.196152] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa81d72f-5fc3-41d2-b8fb-7f107ba4ec9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.202858] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1082.202858] env[61936]: value = "task-1253522" [ 1082.202858] env[61936]: _type = "Task" [ 1082.202858] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.210281] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.277098] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.277098] env[61936]: DEBUG nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1082.279680] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.583s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.281152] env[61936]: INFO nova.compute.claims [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.653953] env[61936]: DEBUG nova.compute.manager [req-b62693d2-207a-400d-8cfe-23dbd0ebd6f3 req-0403c60e-c30c-4324-857c-064772ce58ef service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Received event network-vif-plugged-c0d12433-0c29-4d8f-9664-dcbb76583dfc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1082.654212] env[61936]: DEBUG oslo_concurrency.lockutils [req-b62693d2-207a-400d-8cfe-23dbd0ebd6f3 req-0403c60e-c30c-4324-857c-064772ce58ef service nova] Acquiring lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.654428] env[61936]: DEBUG oslo_concurrency.lockutils [req-b62693d2-207a-400d-8cfe-23dbd0ebd6f3 req-0403c60e-c30c-4324-857c-064772ce58ef service nova] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.654600] env[61936]: DEBUG oslo_concurrency.lockutils [req-b62693d2-207a-400d-8cfe-23dbd0ebd6f3 req-0403c60e-c30c-4324-857c-064772ce58ef service nova] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.654767] env[61936]: DEBUG nova.compute.manager [req-b62693d2-207a-400d-8cfe-23dbd0ebd6f3 req-0403c60e-c30c-4324-857c-064772ce58ef service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] No waiting events found dispatching network-vif-plugged-c0d12433-0c29-4d8f-9664-dcbb76583dfc {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1082.654937] env[61936]: WARNING nova.compute.manager [req-b62693d2-207a-400d-8cfe-23dbd0ebd6f3 req-0403c60e-c30c-4324-857c-064772ce58ef service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Received unexpected event network-vif-plugged-c0d12433-0c29-4d8f-9664-dcbb76583dfc for instance with vm_state building and task_state spawning. [ 1082.713487] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253522, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.780324] env[61936]: DEBUG nova.network.neutron [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Successfully updated port: c0d12433-0c29-4d8f-9664-dcbb76583dfc {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.786481] env[61936]: DEBUG nova.compute.utils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.797570] env[61936]: DEBUG nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1082.797827] env[61936]: DEBUG nova.network.neutron [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1082.848990] env[61936]: DEBUG nova.policy [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1083.123726] env[61936]: DEBUG nova.network.neutron [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Successfully created port: ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1083.213663] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253522, 'name': PowerOnVM_Task, 'duration_secs': 0.607731} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.214055] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.214164] env[61936]: DEBUG nova.compute.manager [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1083.214891] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96860e11-2e76-4914-b628-dc0dc116991f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.285736] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.285855] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.286047] env[61936]: DEBUG nova.network.neutron [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1083.298136] env[61936]: DEBUG nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1083.438403] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9d7fbf-b7a0-45f5-a425-8fcf02b958d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.446651] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98e44a3-aff4-4e3e-a7db-b7196ec53b67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.476695] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c094cbe-674d-4790-80f8-a4dbe791fb12 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.484747] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5a8309-f322-415f-9a18-aecacca0e5d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.500045] env[61936]: DEBUG nova.compute.provider_tree [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.724798] env[61936]: INFO nova.compute.manager [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] bringing vm to original state: 'stopped' [ 1083.818645] env[61936]: DEBUG nova.network.neutron [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1083.941937] env[61936]: DEBUG nova.network.neutron [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance_info_cache with network_info: [{"id": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "address": "fa:16:3e:a3:46:a9", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0d12433-0c", "ovs_interfaceid": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.002538] env[61936]: DEBUG nova.scheduler.client.report [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1084.307333] env[61936]: DEBUG nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1084.331844] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1084.332149] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.332348] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1084.332579] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.332772] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1084.333118] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1084.333399] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1084.333594] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1084.333807] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1084.334016] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1084.334244] env[61936]: DEBUG nova.virt.hardware [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1084.335204] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458df3bf-37d5-4a15-9c46-ae3828986b61 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.343857] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f601ab1c-e782-4e01-a036-7c99529ba285 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.445034] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.445389] env[61936]: DEBUG nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Instance network_info: |[{"id": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "address": "fa:16:3e:a3:46:a9", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0d12433-0c", "ovs_interfaceid": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1084.445860] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:46:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0d12433-0c29-4d8f-9664-dcbb76583dfc', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.453852] env[61936]: DEBUG oslo.service.loopingcall [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.454090] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1084.454370] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0798ccb1-f0a7-4164-a1b3-5960d4ab5cb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.474389] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.474389] env[61936]: value = "task-1253523" [ 1084.474389] env[61936]: _type = "Task" [ 1084.474389] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.483942] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253523, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.508606] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.229s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.509231] env[61936]: DEBUG nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1084.650621] env[61936]: DEBUG nova.network.neutron [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Successfully updated port: ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1084.696415] env[61936]: DEBUG nova.compute.manager [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Received event network-changed-c0d12433-0c29-4d8f-9664-dcbb76583dfc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1084.696705] env[61936]: DEBUG nova.compute.manager [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Refreshing instance network info cache due to event network-changed-c0d12433-0c29-4d8f-9664-dcbb76583dfc. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1084.696943] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Acquiring lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.697116] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Acquired lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.697283] env[61936]: DEBUG nova.network.neutron [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Refreshing network info cache for port c0d12433-0c29-4d8f-9664-dcbb76583dfc {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1084.731260] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.731546] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.731756] env[61936]: DEBUG nova.compute.manager [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1084.735016] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b907be21-2c0a-48d3-aeaf-8fe1c4a8db2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.741335] env[61936]: DEBUG nova.compute.manager [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61936) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1084.984879] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253523, 'name': CreateVM_Task, 'duration_secs': 0.310839} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.984879] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1084.985930] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.985930] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.986091] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1084.986353] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bf9387b-c9c7-4060-a76d-eaacef8b6a62 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.991097] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1084.991097] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e2b399-a124-d5bc-8570-a7770c55f5bf" [ 1084.991097] env[61936]: _type = "Task" [ 1084.991097] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.998774] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e2b399-a124-d5bc-8570-a7770c55f5bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.013659] env[61936]: DEBUG nova.compute.utils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1085.014975] env[61936]: DEBUG nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1085.015157] env[61936]: DEBUG nova.network.neutron [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1085.063895] env[61936]: DEBUG nova.policy [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9260b79bb7f4e44bb1f50b16273df09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ce9d789b30c49758bd073d17b3f1281', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1085.153962] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.154128] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.154282] env[61936]: DEBUG nova.network.neutron [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1085.244973] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1085.245301] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ce158e0-1622-4a94-829d-2112f3f1b513 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.254385] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1085.254385] env[61936]: value = "task-1253524" [ 1085.254385] env[61936]: _type = "Task" [ 1085.254385] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.261323] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.379405] env[61936]: DEBUG nova.network.neutron [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Successfully created port: 92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1085.423142] env[61936]: DEBUG nova.network.neutron [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updated VIF entry in instance network info cache for port c0d12433-0c29-4d8f-9664-dcbb76583dfc. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1085.423552] env[61936]: DEBUG nova.network.neutron [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance_info_cache with network_info: [{"id": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "address": "fa:16:3e:a3:46:a9", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0d12433-0c", "ovs_interfaceid": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.503520] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52e2b399-a124-d5bc-8570-a7770c55f5bf, 'name': SearchDatastore_Task, 'duration_secs': 0.027158} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.504051] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.504399] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.504722] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.504949] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.505347] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.505752] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66a6826d-b1ea-4d0f-b95d-440c35d51dad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.518467] env[61936]: DEBUG nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1085.522955] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.523157] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1085.524115] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad2c2b15-0a9c-4fa2-b921-5caf33afa5e7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.529380] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1085.529380] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528f23a9-13f7-a946-ab22-ab81e6c4b47a" [ 1085.529380] env[61936]: _type = "Task" [ 1085.529380] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.537459] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528f23a9-13f7-a946-ab22-ab81e6c4b47a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.591540] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1085.591757] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1085.591907] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 1085.687851] env[61936]: DEBUG nova.network.neutron [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1085.767022] env[61936]: DEBUG oslo_vmware.api [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253524, 'name': PowerOffVM_Task, 'duration_secs': 0.279066} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.767524] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.767942] env[61936]: DEBUG nova.compute.manager [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1085.768810] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414c8fbb-967f-46d9-90e8-ab257aeac5f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.927802] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Releasing lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.927802] env[61936]: DEBUG nova.compute.manager [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-vif-plugged-ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1085.927802] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Acquiring lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.927802] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.927802] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.927802] env[61936]: DEBUG nova.compute.manager [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] No waiting events found dispatching network-vif-plugged-ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1085.927802] env[61936]: WARNING nova.compute.manager [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received unexpected event network-vif-plugged-ef5dde7f-60d3-40b7-877c-53c75f17e14e for instance with vm_state building and task_state spawning. [ 1085.927802] env[61936]: DEBUG nova.compute.manager [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1085.927802] env[61936]: DEBUG nova.compute.manager [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing instance network info cache due to event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1085.927802] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.022309] env[61936]: DEBUG nova.network.neutron [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.039644] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]528f23a9-13f7-a946-ab22-ab81e6c4b47a, 'name': SearchDatastore_Task, 'duration_secs': 0.012229} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.040463] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-425323d6-1146-46ae-a07e-ea2bedf1eb55 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.047675] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1086.047675] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52364085-fc35-1b99-0500-0bf96f7ac094" [ 1086.047675] env[61936]: _type = "Task" [ 1086.047675] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.055820] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52364085-fc35-1b99-0500-0bf96f7ac094, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.285089] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.553s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.525173] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.525544] env[61936]: DEBUG nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Instance network_info: |[{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1086.525839] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.526018] env[61936]: DEBUG nova.network.neutron [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1086.527151] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:d1:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6ffea58a-e043-4ad1-9ad7-45a61fdca001', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef5dde7f-60d3-40b7-877c-53c75f17e14e', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1086.534777] env[61936]: DEBUG oslo.service.loopingcall [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.536532] env[61936]: DEBUG nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1086.538472] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1086.539517] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03e9f355-dcc2-4ebc-845c-323afed046ed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.561495] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1086.561768] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1086.561919] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1086.562128] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1086.562281] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1086.562429] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1086.562634] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1086.562820] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1086.562955] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1086.563132] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1086.563306] env[61936]: DEBUG nova.virt.hardware [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1086.564289] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5788f918-791a-490d-b201-20eb7e61fd85 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.570768] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.570768] env[61936]: value = "task-1253525" [ 1086.570768] env[61936]: _type = "Task" [ 1086.570768] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.571009] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52364085-fc35-1b99-0500-0bf96f7ac094, 'name': SearchDatastore_Task, 'duration_secs': 0.009257} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.571626] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.571930] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 6c78d80f-84e8-4250-a60a-c9e28befd07a/6c78d80f-84e8-4250-a60a-c9e28befd07a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1086.574835] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f73b4c8a-aaef-442e-948d-046e6429cde1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.579899] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0d5fcd-e648-4f93-aa23-dc6297e90378 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.587374] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253525, 'name': CreateVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.587742] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1086.587742] env[61936]: value = "task-1253526" [ 1086.587742] env[61936]: _type = "Task" [ 1086.587742] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.605354] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.766116] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.768474] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.768726] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.768964] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.769166] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.771609] env[61936]: INFO nova.compute.manager [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Terminating instance [ 1086.793893] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.793893] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.793893] env[61936]: DEBUG nova.objects.instance [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1087.061303] env[61936]: DEBUG nova.compute.manager [req-91b5311d-cb09-4c95-ac5a-eb73cc18ca7f req-7ffd78af-9465-4fdd-9335-a97e7404b883 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Received event network-vif-plugged-92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1087.061303] env[61936]: DEBUG oslo_concurrency.lockutils [req-91b5311d-cb09-4c95-ac5a-eb73cc18ca7f req-7ffd78af-9465-4fdd-9335-a97e7404b883 service nova] Acquiring lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.061303] env[61936]: DEBUG oslo_concurrency.lockutils [req-91b5311d-cb09-4c95-ac5a-eb73cc18ca7f req-7ffd78af-9465-4fdd-9335-a97e7404b883 service nova] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.061303] env[61936]: DEBUG oslo_concurrency.lockutils [req-91b5311d-cb09-4c95-ac5a-eb73cc18ca7f req-7ffd78af-9465-4fdd-9335-a97e7404b883 service nova] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.061616] env[61936]: DEBUG nova.compute.manager [req-91b5311d-cb09-4c95-ac5a-eb73cc18ca7f req-7ffd78af-9465-4fdd-9335-a97e7404b883 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] No waiting events found dispatching network-vif-plugged-92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1087.061897] env[61936]: WARNING nova.compute.manager [req-91b5311d-cb09-4c95-ac5a-eb73cc18ca7f req-7ffd78af-9465-4fdd-9335-a97e7404b883 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Received unexpected event network-vif-plugged-92b2f7ae-8a1a-4908-939d-f2b01d922c5f for instance with vm_state building and task_state spawning. [ 1087.084277] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253525, 'name': CreateVM_Task, 'duration_secs': 0.493829} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.084497] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1087.085116] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.086021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.086021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1087.086021] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8184eb6c-3585-4776-af22-522950c64eeb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.092578] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1087.092578] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52184ab8-66b5-dfbf-4251-b24fe48b7f4d" [ 1087.092578] env[61936]: _type = "Task" [ 1087.092578] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.099307] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490775} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.099834] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 6c78d80f-84e8-4250-a60a-c9e28befd07a/6c78d80f-84e8-4250-a60a-c9e28befd07a.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1087.100071] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.100326] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01608baf-7fdf-4469-9467-47fe602aa506 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.106324] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52184ab8-66b5-dfbf-4251-b24fe48b7f4d, 'name': SearchDatastore_Task, 'duration_secs': 0.00743} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.106882] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.107137] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.107889] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.107889] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.107889] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.108112] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69e9bd8a-463a-4647-a40d-fa3f3b1894e9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.110643] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1087.110643] env[61936]: value = "task-1253527" [ 1087.110643] env[61936]: _type = "Task" [ 1087.110643] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.121030] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.126021] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.126021] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1087.126021] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a88d2f6-c66d-4b63-9c16-fb13d642aea8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.128791] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1087.128791] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ce39b6-cd18-e4a3-bbc1-2b60aee76e1d" [ 1087.128791] env[61936]: _type = "Task" [ 1087.128791] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.139383] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ce39b6-cd18-e4a3-bbc1-2b60aee76e1d, 'name': SearchDatastore_Task, 'duration_secs': 0.007587} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.140152] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9d40489-0469-41ee-bf0b-94fb8f56169c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.145391] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1087.145391] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5271261a-ac97-dc51-4d5d-d5aeb35f165a" [ 1087.145391] env[61936]: _type = "Task" [ 1087.145391] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.153140] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5271261a-ac97-dc51-4d5d-d5aeb35f165a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.162643] env[61936]: DEBUG nova.network.neutron [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Successfully updated port: 92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1087.275976] env[61936]: DEBUG nova.compute.manager [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1087.276122] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1087.277061] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71db0982-65d9-4003-81c0-b7701ab66514 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.285051] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1087.285300] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4359a95-4cad-4779-800c-be2f6ef255b0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.295442] env[61936]: DEBUG nova.network.neutron [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updated VIF entry in instance network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1087.295530] env[61936]: DEBUG nova.network.neutron [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.342599] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1087.342853] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1087.343018] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleting the datastore file [datastore1] 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.343295] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82173881-68c3-4af7-9a9e-54f576ba9e40 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.352582] env[61936]: DEBUG oslo_vmware.api [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1087.352582] env[61936]: value = "task-1253529" [ 1087.352582] env[61936]: _type = "Task" [ 1087.352582] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.360738] env[61936]: DEBUG oslo_vmware.api [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253529, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.623728] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058238} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.624088] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1087.624796] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2cb564-3c58-4e07-a027-1f362f53f5ec {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.646626] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 6c78d80f-84e8-4250-a60a-c9e28befd07a/6c78d80f-84e8-4250-a60a-c9e28befd07a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.647129] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-def0ad9f-5b09-4b9f-864d-ec3f0820b90d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.665305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.665429] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.665603] env[61936]: DEBUG nova.network.neutron [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.669639] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5271261a-ac97-dc51-4d5d-d5aeb35f165a, 'name': SearchDatastore_Task, 'duration_secs': 0.007141} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.670897] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.671172] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] ea74d4a3-556f-4290-93ea-7f55f8c8bc2c/ea74d4a3-556f-4290-93ea-7f55f8c8bc2c.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1087.671476] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1087.671476] env[61936]: value = "task-1253530" [ 1087.671476] env[61936]: _type = "Task" [ 1087.671476] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.671654] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e26ceb3d-ac65-4ed3-827f-0a32d0f7e196 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.682965] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253530, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.684313] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1087.684313] env[61936]: value = "task-1253531" [ 1087.684313] env[61936]: _type = "Task" [ 1087.684313] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.691501] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.800614] env[61936]: DEBUG oslo_concurrency.lockutils [req-1f76f6a2-4ba9-4612-b5f8-30d28fd00a1a req-fb587cde-671c-44ec-92d4-0fa8657eaae1 service nova] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.801860] env[61936]: DEBUG oslo_concurrency.lockutils [None req-bbe5a00c-4872-4642-b77d-dde9b90ead18 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.863440] env[61936]: DEBUG oslo_vmware.api [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253529, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123691} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.863713] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1087.863903] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1087.864098] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1087.864312] env[61936]: INFO nova.compute.manager [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1087.864573] env[61936]: DEBUG oslo.service.loopingcall [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1087.864772] env[61936]: DEBUG nova.compute.manager [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1087.864877] env[61936]: DEBUG nova.network.neutron [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1088.186986] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253530, 'name': ReconfigVM_Task, 'duration_secs': 0.508279} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.192250] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 6c78d80f-84e8-4250-a60a-c9e28befd07a/6c78d80f-84e8-4250-a60a-c9e28befd07a.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.192250] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fc4b1eb-a648-430d-837d-e41875a5b3ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.200210] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431501} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.201553] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] ea74d4a3-556f-4290-93ea-7f55f8c8bc2c/ea74d4a3-556f-4290-93ea-7f55f8c8bc2c.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1088.201678] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1088.202074] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1088.202074] env[61936]: value = "task-1253532" [ 1088.202074] env[61936]: _type = "Task" [ 1088.202074] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.202271] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bddd1e3a-ec7d-4593-9111-fdd1f89ba206 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.214090] env[61936]: DEBUG nova.network.neutron [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1088.222931] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253532, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.224299] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1088.224299] env[61936]: value = "task-1253533" [ 1088.224299] env[61936]: _type = "Task" [ 1088.224299] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.226257] env[61936]: DEBUG nova.compute.manager [req-17f71122-d578-48d8-ae29-ca6ecbd17b2e req-b6f7b478-b54e-4522-88ff-021ea6d10179 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Received event network-vif-deleted-40113568-3584-481c-aeb0-65d87745501c {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1088.226448] env[61936]: INFO nova.compute.manager [req-17f71122-d578-48d8-ae29-ca6ecbd17b2e req-b6f7b478-b54e-4522-88ff-021ea6d10179 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Neutron deleted interface 40113568-3584-481c-aeb0-65d87745501c; detaching it from the instance and deleting it from the info cache [ 1088.226611] env[61936]: DEBUG nova.network.neutron [req-17f71122-d578-48d8-ae29-ca6ecbd17b2e req-b6f7b478-b54e-4522-88ff-021ea6d10179 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.236420] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253533, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.355835] env[61936]: DEBUG nova.network.neutron [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updating instance_info_cache with network_info: [{"id": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "address": "fa:16:3e:90:92:fb", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b2f7ae-8a", "ovs_interfaceid": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.625497] env[61936]: DEBUG nova.network.neutron [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.718104] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253532, 'name': Rename_Task, 'duration_secs': 0.317561} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.718395] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.718646] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-deed2119-ffeb-4760-a4a5-8e819361ac35 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.725112] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1088.725112] env[61936]: value = "task-1253534" [ 1088.725112] env[61936]: _type = "Task" [ 1088.725112] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.735126] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-626773af-e69c-4a2b-8a5b-258f95144868 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.737016] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253533, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063409} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.739868] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1088.740159] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253534, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.741170] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbecc17-278b-4835-bef5-14dfe3e47ee3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.746946] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebb0641-de79-4429-9856-4f0b003bde88 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.775116] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] ea74d4a3-556f-4290-93ea-7f55f8c8bc2c/ea74d4a3-556f-4290-93ea-7f55f8c8bc2c.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.775797] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a76f2d4-84ee-45b0-b862-1d6973a16173 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.798123] env[61936]: DEBUG nova.compute.manager [req-17f71122-d578-48d8-ae29-ca6ecbd17b2e req-b6f7b478-b54e-4522-88ff-021ea6d10179 service nova] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Detach interface failed, port_id=40113568-3584-481c-aeb0-65d87745501c, reason: Instance 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1088.803774] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1088.803774] env[61936]: value = "task-1253535" [ 1088.803774] env[61936]: _type = "Task" [ 1088.803774] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.811164] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253535, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.858908] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.859249] env[61936]: DEBUG nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Instance network_info: |[{"id": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "address": "fa:16:3e:90:92:fb", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b2f7ae-8a", "ovs_interfaceid": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1088.859716] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:92:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92b2f7ae-8a1a-4908-939d-f2b01d922c5f', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1088.867227] env[61936]: DEBUG oslo.service.loopingcall [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.867440] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1088.867663] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f390d151-e090-42ff-941f-631aed292b13 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.886821] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1088.886821] env[61936]: value = "task-1253536" [ 1088.886821] env[61936]: _type = "Task" [ 1088.886821] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.894184] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253536, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.091532] env[61936]: DEBUG nova.compute.manager [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Received event network-changed-92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1089.091658] env[61936]: DEBUG nova.compute.manager [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Refreshing instance network info cache due to event network-changed-92b2f7ae-8a1a-4908-939d-f2b01d922c5f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1089.091926] env[61936]: DEBUG oslo_concurrency.lockutils [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] Acquiring lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.092070] env[61936]: DEBUG oslo_concurrency.lockutils [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] Acquired lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.092235] env[61936]: DEBUG nova.network.neutron [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Refreshing network info cache for port 92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.127599] env[61936]: INFO nova.compute.manager [-] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Took 1.26 seconds to deallocate network for instance. [ 1089.237831] env[61936]: DEBUG oslo_vmware.api [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253534, 'name': PowerOnVM_Task, 'duration_secs': 0.464676} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.238228] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1089.238499] env[61936]: INFO nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1089.238755] env[61936]: DEBUG nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1089.239577] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5624816-54d3-4ed3-8253-91741255c372 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.312901] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253535, 'name': ReconfigVM_Task, 'duration_secs': 0.32466} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.313232] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Reconfigured VM instance instance-0000006a to attach disk [datastore1] ea74d4a3-556f-4290-93ea-7f55f8c8bc2c/ea74d4a3-556f-4290-93ea-7f55f8c8bc2c.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.313858] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33b45e73-7126-4b7a-a31b-efe2b2313b82 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.320375] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1089.320375] env[61936]: value = "task-1253537" [ 1089.320375] env[61936]: _type = "Task" [ 1089.320375] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.328216] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253537, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.396363] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253536, 'name': CreateVM_Task, 'duration_secs': 0.346951} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.396534] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1089.397236] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.397418] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.397754] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1089.398012] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-073aa93d-820c-44ae-82b1-cf83d7806d9b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.402165] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1089.402165] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5279591f-39c6-abe5-6fb2-edd301f85136" [ 1089.402165] env[61936]: _type = "Task" [ 1089.402165] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.409162] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5279591f-39c6-abe5-6fb2-edd301f85136, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.633850] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.634179] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.634351] env[61936]: DEBUG nova.objects.instance [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'resources' on Instance uuid 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.758432] env[61936]: INFO nova.compute.manager [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Took 13.99 seconds to build instance. [ 1089.803330] env[61936]: DEBUG nova.network.neutron [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updated VIF entry in instance network info cache for port 92b2f7ae-8a1a-4908-939d-f2b01d922c5f. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1089.803687] env[61936]: DEBUG nova.network.neutron [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updating instance_info_cache with network_info: [{"id": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "address": "fa:16:3e:90:92:fb", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b2f7ae-8a", "ovs_interfaceid": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.831586] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253537, 'name': Rename_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.912599] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5279591f-39c6-abe5-6fb2-edd301f85136, 'name': SearchDatastore_Task, 'duration_secs': 0.089967} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.912988] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.913293] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1089.913603] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.913940] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.914021] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1089.914327] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0dfc2dbc-6caf-4795-bc83-db43cc2a28a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.921867] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1089.922065] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1089.922765] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8335b74-df21-48f2-8dca-a4105dd278bd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.928198] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1089.928198] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d8236a-c564-8c7d-0964-85898ba2bb3f" [ 1089.928198] env[61936]: _type = "Task" [ 1089.928198] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.936518] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d8236a-c564-8c7d-0964-85898ba2bb3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.151722] env[61936]: DEBUG nova.scheduler.client.report [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Refreshing inventories for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1090.166218] env[61936]: DEBUG nova.scheduler.client.report [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating ProviderTree inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1090.166465] env[61936]: DEBUG nova.compute.provider_tree [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1090.177381] env[61936]: DEBUG nova.scheduler.client.report [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Refreshing aggregate associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, aggregates: None {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1090.192823] env[61936]: DEBUG nova.scheduler.client.report [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Refreshing trait associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1090.259682] env[61936]: DEBUG oslo_concurrency.lockutils [None req-11dccbff-4388-4930-9ef9-a7a4c4372b97 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.505s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.306331] env[61936]: DEBUG oslo_concurrency.lockutils [req-3820be88-3dd3-4d02-bfe4-986dc63d93c9 req-b494f9aa-d615-4c51-8ed4-7b87b0abf296 service nova] Releasing lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.320993] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5b8168-39b0-4555-8cc2-6fcededd4adb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.333167] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253537, 'name': Rename_Task, 'duration_secs': 0.554613} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.334897] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1090.335190] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c533271c-3f82-4de0-b922-99430097ef17 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.337186] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282ada80-bf26-4be5-aa20-d08562ba38de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.368785] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d467582-8936-42df-a661-e5b02bd33c21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.371665] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1090.371665] env[61936]: value = "task-1253538" [ 1090.371665] env[61936]: _type = "Task" [ 1090.371665] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.378475] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5114bad3-bf2a-4407-9dd6-ec0f6b099090 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.385317] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253538, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.395082] env[61936]: DEBUG nova.compute.provider_tree [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1090.440185] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52d8236a-c564-8c7d-0964-85898ba2bb3f, 'name': SearchDatastore_Task, 'duration_secs': 0.009101} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.441080] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27d89f44-1a84-46e0-b58d-19b30d0b9db1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.447572] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1090.447572] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52fa5fba-2d0e-f8ab-cedb-e6ece07b0af7" [ 1090.447572] env[61936]: _type = "Task" [ 1090.447572] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.455896] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52fa5fba-2d0e-f8ab-cedb-e6ece07b0af7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.882757] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253538, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.898391] env[61936]: DEBUG nova.scheduler.client.report [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1090.957502] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52fa5fba-2d0e-f8ab-cedb-e6ece07b0af7, 'name': SearchDatastore_Task, 'duration_secs': 0.014236} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.957785] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.958076] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] b8e0c45d-341e-4ad4-8941-4532ebc53e40/b8e0c45d-341e-4ad4-8941-4532ebc53e40.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1090.958352] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-19bff43f-880e-44b6-8382-5946369d98ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.965269] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1090.965269] env[61936]: value = "task-1253539" [ 1090.965269] env[61936]: _type = "Task" [ 1090.965269] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.974688] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253539, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.977136] env[61936]: DEBUG nova.compute.manager [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Stashing vm_state: active {{(pid=61936) _prep_resize /opt/stack/nova/nova/compute/manager.py:6015}} [ 1091.126840] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Didn't find any instances for network info cache update. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10412}} [ 1091.127105] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.127351] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.127551] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.127742] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.127915] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.128115] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.128290] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 1091.128452] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.385239] env[61936]: DEBUG oslo_vmware.api [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253538, 'name': PowerOnVM_Task, 'duration_secs': 0.544942} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.385479] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1091.385676] env[61936]: INFO nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Took 7.08 seconds to spawn the instance on the hypervisor. [ 1091.385857] env[61936]: DEBUG nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1091.386650] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e98d21-717f-4cb8-bbee-233e0e72b8cb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.403505] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.769s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.426819] env[61936]: INFO nova.scheduler.client.report [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted allocations for instance 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc [ 1091.475869] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253539, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.433317} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.476111] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] b8e0c45d-341e-4ad4-8941-4532ebc53e40/b8e0c45d-341e-4ad4-8941-4532ebc53e40.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1091.476248] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.476459] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0693ad57-bf42-4b92-b123-71ef3e9816b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.485196] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1091.485196] env[61936]: value = "task-1253540" [ 1091.485196] env[61936]: _type = "Task" [ 1091.485196] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.493594] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.497464] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.497711] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.632572] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.905736] env[61936]: INFO nova.compute.manager [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Took 15.54 seconds to build instance. [ 1091.934448] env[61936]: DEBUG oslo_concurrency.lockutils [None req-023122db-2ff7-41d5-b8cd-bfa4f8b2862b tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.166s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.995152] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24081} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.995493] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1091.996368] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40446551-ce8e-408a-8561-be787d132cb3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.003246] env[61936]: INFO nova.compute.claims [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1092.028348] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] b8e0c45d-341e-4ad4-8941-4532ebc53e40/b8e0c45d-341e-4ad4-8941-4532ebc53e40.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.029422] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82491fc4-12c6-4629-8967-c9e3240c989a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.051595] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1092.051595] env[61936]: value = "task-1253541" [ 1092.051595] env[61936]: _type = "Task" [ 1092.051595] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.060840] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.408238] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a0c4dc27-c43e-455e-91eb-5bbb4678ded4 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.045s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.511049] env[61936]: INFO nova.compute.resource_tracker [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating resource usage from migration 0b3d90bf-182f-478f-bd29-db3c331088fe [ 1092.568170] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253541, 'name': ReconfigVM_Task, 'duration_secs': 0.293988} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.568710] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Reconfigured VM instance instance-0000006b to attach disk [datastore1] b8e0c45d-341e-4ad4-8941-4532ebc53e40/b8e0c45d-341e-4ad4-8941-4532ebc53e40.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.570016] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-486baa2f-9eb9-4f4d-9007-28ccc508ca89 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.578144] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1092.578144] env[61936]: value = "task-1253542" [ 1092.578144] env[61936]: _type = "Task" [ 1092.578144] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.590187] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253542, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.595213] env[61936]: DEBUG nova.compute.manager [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1092.595463] env[61936]: DEBUG nova.compute.manager [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing instance network info cache due to event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1092.595606] env[61936]: DEBUG oslo_concurrency.lockutils [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.595985] env[61936]: DEBUG oslo_concurrency.lockutils [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.596159] env[61936]: DEBUG nova.network.neutron [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.706256] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97fa9b7-e701-4db6-82ee-f4f140a94913 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.715670] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96aa893-a54e-4fc2-a83a-ea80efaca310 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.748752] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d00b2f0-19b0-41a1-b55e-05455f05c71e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.756742] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967b0a7a-528d-415d-9bfd-e73dd5ac576d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.770847] env[61936]: DEBUG nova.compute.provider_tree [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.089256] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253542, 'name': Rename_Task, 'duration_secs': 0.242766} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.089256] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1093.089256] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a5a0945-b6d2-4153-a29a-669178bf8a6a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.094750] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1093.094750] env[61936]: value = "task-1253543" [ 1093.094750] env[61936]: _type = "Task" [ 1093.094750] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.103390] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.165231] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "484e5d41-29f4-4845-9633-157c03766978" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.165472] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.273636] env[61936]: DEBUG nova.scheduler.client.report [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1093.363225] env[61936]: DEBUG nova.network.neutron [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updated VIF entry in instance network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1093.363703] env[61936]: DEBUG nova.network.neutron [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.605730] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253543, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.669111] env[61936]: INFO nova.compute.manager [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Detaching volume 0daa7ba3-e820-4de6-a26e-a0fecebe9746 [ 1093.708466] env[61936]: INFO nova.virt.block_device [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Attempting to driver detach volume 0daa7ba3-e820-4de6-a26e-a0fecebe9746 from mountpoint /dev/sdb [ 1093.708690] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1093.708927] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270028', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'name': 'volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '484e5d41-29f4-4845-9633-157c03766978', 'attached_at': '', 'detached_at': '', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'serial': '0daa7ba3-e820-4de6-a26e-a0fecebe9746'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1093.710602] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d91608-077a-4b48-888a-f62ec708f7d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.733592] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7815a3b4-d97e-407c-aead-8b9d7f283b2d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.741600] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76886168-1a9e-4cc7-8985-dab132385aba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.765418] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286af3fc-97a5-4b57-831d-1771b0fa5017 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.781438] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.283s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.781550] env[61936]: INFO nova.compute.manager [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Migrating [ 1093.788066] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] The volume has not been displaced from its original location: [datastore1] volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746/volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1093.793408] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1093.796305] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.164s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.796492] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.796639] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1093.796898] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2dedc03-cc20-4bec-8918-ebbcbbdb04b3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.817236] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8a67ee-9d63-48a4-b559-4557a5faeaad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.824749] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb76a46-1441-4fd3-8c99-e506f0651eb4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.830680] env[61936]: DEBUG oslo_vmware.api [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1093.830680] env[61936]: value = "task-1253544" [ 1093.830680] env[61936]: _type = "Task" [ 1093.830680] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.845089] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11058fb2-c6be-47b9-8f53-c9d97b59c4ee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.852259] env[61936]: DEBUG oslo_vmware.api [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253544, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.856852] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067f79c2-6ccc-4bfb-b647-0becb4b5f9b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.889139] env[61936]: DEBUG oslo_concurrency.lockutils [req-3ba7742e-0c18-40c6-855c-26137da572e2 req-37ad82f9-6404-4d09-80b3-f5e368527b73 service nova] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.889714] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180035MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1093.889851] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.890062] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.106126] env[61936]: DEBUG oslo_vmware.api [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253543, 'name': PowerOnVM_Task, 'duration_secs': 0.639077} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.106553] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1094.106906] env[61936]: INFO nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Took 7.57 seconds to spawn the instance on the hypervisor. [ 1094.107235] env[61936]: DEBUG nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1094.108146] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fff9d26-73f5-4104-b74a-1b9661f995a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.322378] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.322378] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.322378] env[61936]: DEBUG nova.network.neutron [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.343052] env[61936]: DEBUG oslo_vmware.api [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253544, 'name': ReconfigVM_Task, 'duration_secs': 0.434774} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.343822] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1094.348456] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd0e11d7-ba22-4a9a-abdc-6a1fdd013404 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.366865] env[61936]: DEBUG oslo_vmware.api [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1094.366865] env[61936]: value = "task-1253545" [ 1094.366865] env[61936]: _type = "Task" [ 1094.366865] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.374969] env[61936]: DEBUG oslo_vmware.api [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253545, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.400849] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.400849] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.626801] env[61936]: DEBUG nova.compute.manager [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1094.627046] env[61936]: DEBUG nova.compute.manager [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing instance network info cache due to event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1094.627224] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.627366] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.627537] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1094.632669] env[61936]: INFO nova.compute.manager [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Took 14.96 seconds to build instance. [ 1094.875274] env[61936]: DEBUG oslo_vmware.api [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253545, 'name': ReconfigVM_Task, 'duration_secs': 0.227748} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.875596] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270028', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'name': 'volume-0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '484e5d41-29f4-4845-9633-157c03766978', 'attached_at': '', 'detached_at': '', 'volume_id': '0daa7ba3-e820-4de6-a26e-a0fecebe9746', 'serial': '0daa7ba3-e820-4de6-a26e-a0fecebe9746'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1094.902808] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Applying migration context for instance 6c78d80f-84e8-4250-a60a-c9e28befd07a as it has an incoming, in-progress migration 0b3d90bf-182f-478f-bd29-db3c331088fe. Migration status is pre-migrating {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1094.904159] env[61936]: INFO nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating resource usage from migration 0b3d90bf-182f-478f-bd29-db3c331088fe [ 1094.906434] env[61936]: DEBUG nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1094.928208] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fa845f8e-957c-4c0b-a647-190c32989dcd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.928354] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f3f2a0a8-cfe3-462f-a940-05580d5da32d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.928476] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.928592] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 484e5d41-29f4-4845-9633-157c03766978 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.928707] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e6930eaf-bb63-4553-8cd2-f2eb62fca510 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.928841] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance fbde880a-b47c-49e8-b84d-0efee21d62c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.928980] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f1d45dcb-768d-447e-be49-ef43ad98fa61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.929095] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance ea74d4a3-556f-4290-93ea-7f55f8c8bc2c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.929210] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance b8e0c45d-341e-4ad4-8941-4532ebc53e40 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1094.929322] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Migration 0b3d90bf-182f-478f-bd29-db3c331088fe is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1094.929432] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 6c78d80f-84e8-4250-a60a-c9e28befd07a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1095.021179] env[61936]: DEBUG nova.network.neutron [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance_info_cache with network_info: [{"id": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "address": "fa:16:3e:a3:46:a9", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0d12433-0c", "ovs_interfaceid": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.134088] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f0e31787-27ed-45b9-8b9f-e4b60114794b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.464s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.338584] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updated VIF entry in instance network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1095.338943] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.432383] env[61936]: DEBUG nova.objects.instance [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lazy-loading 'flavor' on Instance uuid 484e5d41-29f4-4845-9633-157c03766978 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.435851] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 3e337ad6-194b-4cd5-bfee-708ac00b7430 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1095.435851] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1095.435851] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1095.439618] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.524051] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.613354] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407b91a1-29c4-42b8-9f6f-420d66971dad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.620972] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50256a2-d17c-4e75-9ddb-a96ae71cc173 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.650308] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ab4f40-84a6-4457-bf39-c18c6f4c49c6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.657193] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3367ff2-01cb-4bb7-b0b9-d86221c5991c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.671327] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.842135] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.842417] env[61936]: DEBUG nova.compute.manager [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1095.842589] env[61936]: DEBUG nova.compute.manager [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing instance network info cache due to event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1095.842799] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.842941] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.843121] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1095.992468] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.992468] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.174606] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.440282] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ad87abd2-5224-4d0d-be8a-286f18287e07 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.275s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.494859] env[61936]: INFO nova.compute.manager [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Detaching volume 41f3c9d7-6e22-4c69-9d06-47b73b138c6e [ 1096.529453] env[61936]: INFO nova.virt.block_device [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Attempting to driver detach volume 41f3c9d7-6e22-4c69-9d06-47b73b138c6e from mountpoint /dev/sdb [ 1096.529750] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1096.529953] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270025', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'name': 'volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fbde880a-b47c-49e8-b84d-0efee21d62c0', 'attached_at': '', 'detached_at': '', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'serial': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1096.530865] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0117ae5-b747-4bef-bad8-3d1598910ba5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.561139] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updated VIF entry in instance network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1096.561481] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.564038] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e339c04e-ce0f-4193-be2d-4c62c6b92144 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.571862] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b4671c-1e99-4d7c-9bfa-31f6138e5a1b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.595688] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94af201-ddfc-40f3-b221-36144967b044 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.613727] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] The volume has not been displaced from its original location: [datastore1] volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e/volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1096.619214] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfiguring VM instance instance-00000055 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1096.619555] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5879be96-3806-422c-a812-08d0e4396166 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.638352] env[61936]: DEBUG oslo_vmware.api [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1096.638352] env[61936]: value = "task-1253546" [ 1096.638352] env[61936]: _type = "Task" [ 1096.638352] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.646254] env[61936]: DEBUG oslo_vmware.api [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253546, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.648569] env[61936]: DEBUG nova.compute.manager [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Received event network-changed-92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1096.648716] env[61936]: DEBUG nova.compute.manager [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Refreshing instance network info cache due to event network-changed-92b2f7ae-8a1a-4908-939d-f2b01d922c5f. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1096.648939] env[61936]: DEBUG oslo_concurrency.lockutils [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] Acquiring lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.649110] env[61936]: DEBUG oslo_concurrency.lockutils [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] Acquired lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.649276] env[61936]: DEBUG nova.network.neutron [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Refreshing network info cache for port 92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1096.679613] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1096.679833] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.790s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.680088] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.241s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.681834] env[61936]: INFO nova.compute.claims [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.044044] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f568e03-a31f-49a9-a637-a5eb77d922e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.063238] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance '6c78d80f-84e8-4250-a60a-c9e28befd07a' progress to 0 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1097.067178] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.067431] env[61936]: DEBUG nova.compute.manager [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1097.067602] env[61936]: DEBUG nova.compute.manager [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing instance network info cache due to event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1097.067802] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.067943] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.068118] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1097.148790] env[61936]: DEBUG oslo_vmware.api [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253546, 'name': ReconfigVM_Task, 'duration_secs': 0.240258} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.149160] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Reconfigured VM instance instance-00000055 to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1097.155591] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fab16f9c-d6d7-40f6-9b92-82d671b69a40 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.170559] env[61936]: DEBUG oslo_vmware.api [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1097.170559] env[61936]: value = "task-1253547" [ 1097.170559] env[61936]: _type = "Task" [ 1097.170559] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.178376] env[61936]: DEBUG oslo_vmware.api [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253547, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.375270] env[61936]: DEBUG nova.network.neutron [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updated VIF entry in instance network info cache for port 92b2f7ae-8a1a-4908-939d-f2b01d922c5f. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1097.375760] env[61936]: DEBUG nova.network.neutron [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updating instance_info_cache with network_info: [{"id": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "address": "fa:16:3e:90:92:fb", "network": {"id": "013cab7b-6fe3-41f8-986a-a5416eb9ab1f", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1329393044-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ce9d789b30c49758bd073d17b3f1281", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b2f7ae-8a", "ovs_interfaceid": "92b2f7ae-8a1a-4908-939d-f2b01d922c5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.532439] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "484e5d41-29f4-4845-9633-157c03766978" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.532708] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.532921] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "484e5d41-29f4-4845-9633-157c03766978-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.533122] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.533329] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.535470] env[61936]: INFO nova.compute.manager [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Terminating instance [ 1097.571295] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.573667] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42e8276e-9099-49d8-934b-e03a5dc464bb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.582757] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1097.582757] env[61936]: value = "task-1253548" [ 1097.582757] env[61936]: _type = "Task" [ 1097.582757] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.591928] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.684647] env[61936]: DEBUG oslo_vmware.api [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253547, 'name': ReconfigVM_Task, 'duration_secs': 0.154086} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.684918] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270025', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'name': 'volume-41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'fbde880a-b47c-49e8-b84d-0efee21d62c0', 'attached_at': '', 'detached_at': '', 'volume_id': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e', 'serial': '41f3c9d7-6e22-4c69-9d06-47b73b138c6e'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1097.799951] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updated VIF entry in instance network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1097.800365] env[61936]: DEBUG nova.network.neutron [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.848973] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efccb170-04bb-4410-bdae-95b4248f8ac5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.856585] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b13838e-20ad-413f-9506-4bf45abbbfe1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.885834] env[61936]: DEBUG oslo_concurrency.lockutils [req-02f1b693-a47f-498f-ae2f-9f55937c9ce0 req-b3ff2d05-f302-4e12-b9f6-15d51369c116 service nova] Releasing lock "refresh_cache-b8e0c45d-341e-4ad4-8941-4532ebc53e40" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.887261] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506cffa3-6dc9-4c39-a1e9-18a49d809b6d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.894612] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2390a5-eccb-4bd8-9c80-ad93bf9f2ee4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.907502] env[61936]: DEBUG nova.compute.provider_tree [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.039066] env[61936]: DEBUG nova.compute.manager [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1098.039325] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1098.040229] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15d36b0-17ce-4300-ac3a-ae47db87915e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.047606] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1098.047849] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9d7334d-5567-4316-9c7f-e53af1e6ca4a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.053165] env[61936]: DEBUG oslo_vmware.api [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1098.053165] env[61936]: value = "task-1253549" [ 1098.053165] env[61936]: _type = "Task" [ 1098.053165] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.061873] env[61936]: DEBUG oslo_vmware.api [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253549, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.092036] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253548, 'name': PowerOffVM_Task, 'duration_secs': 0.357521} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.092208] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.092463] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance '6c78d80f-84e8-4250-a60a-c9e28befd07a' progress to 17 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1098.230851] env[61936]: DEBUG nova.objects.instance [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'flavor' on Instance uuid fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.303389] env[61936]: DEBUG oslo_concurrency.lockutils [req-df19b679-50ee-4636-8112-c5a549272940 req-9bc96be3-a5cf-4a26-ac66-eb69631e64a8 service nova] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.410497] env[61936]: DEBUG nova.scheduler.client.report [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1098.563828] env[61936]: DEBUG oslo_vmware.api [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253549, 'name': PowerOffVM_Task, 'duration_secs': 0.183531} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.564130] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.564310] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.564611] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fde25d06-82b9-4d81-a073-072684ba33ae {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.599818] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1098.599926] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1098.600764] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1098.600764] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1098.600764] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1098.600764] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1098.601412] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1098.601412] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1098.601628] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1098.601817] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1098.602098] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1098.609749] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00ed3a27-9aca-4423-956c-64d36e178617 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.629723] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.630037] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.630321] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleting the datastore file [datastore1] 484e5d41-29f4-4845-9633-157c03766978 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.630664] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8312e09d-39a4-494e-a72c-1fdfc846a757 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.634634] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1098.634634] env[61936]: value = "task-1253551" [ 1098.634634] env[61936]: _type = "Task" [ 1098.634634] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.641020] env[61936]: DEBUG oslo_vmware.api [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1098.641020] env[61936]: value = "task-1253552" [ 1098.641020] env[61936]: _type = "Task" [ 1098.641020] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.649981] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253551, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.655144] env[61936]: DEBUG oslo_vmware.api [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253552, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.915774] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.916326] env[61936]: DEBUG nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1099.144785] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253551, 'name': ReconfigVM_Task, 'duration_secs': 0.185798} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.147767] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance '6c78d80f-84e8-4250-a60a-c9e28befd07a' progress to 33 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1099.155488] env[61936]: DEBUG oslo_vmware.api [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253552, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219545} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.155698] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1099.155880] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1099.156068] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1099.156248] env[61936]: INFO nova.compute.manager [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: 484e5d41-29f4-4845-9633-157c03766978] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1099.156479] env[61936]: DEBUG oslo.service.loopingcall [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.156665] env[61936]: DEBUG nova.compute.manager [-] [instance: 484e5d41-29f4-4845-9633-157c03766978] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1099.156757] env[61936]: DEBUG nova.network.neutron [-] [instance: 484e5d41-29f4-4845-9633-157c03766978] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1099.238241] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b5fbdcb9-601f-4971-94d7-49d2fc759d40 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.246s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.421194] env[61936]: DEBUG nova.compute.utils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1099.422627] env[61936]: DEBUG nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1099.422792] env[61936]: DEBUG nova.network.neutron [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1099.470353] env[61936]: DEBUG nova.policy [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f378ef8b8b5f4ae59c0d71dd1661bb59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa564b684410493fa0028fd345048e02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1099.616784] env[61936]: DEBUG nova.compute.manager [req-b86fac58-2f4d-42a7-9670-a85087041511 req-6dbb4c49-e2fc-4756-a3f8-24fb7ce9d0a8 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Received event network-vif-deleted-7450de29-76d5-40b0-ae76-a79b3455a9bc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1099.617122] env[61936]: INFO nova.compute.manager [req-b86fac58-2f4d-42a7-9670-a85087041511 req-6dbb4c49-e2fc-4756-a3f8-24fb7ce9d0a8 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Neutron deleted interface 7450de29-76d5-40b0-ae76-a79b3455a9bc; detaching it from the instance and deleting it from the info cache [ 1099.617319] env[61936]: DEBUG nova.network.neutron [req-b86fac58-2f4d-42a7-9670-a85087041511 req-6dbb4c49-e2fc-4756-a3f8-24fb7ce9d0a8 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.656565] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1099.656565] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1099.656565] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1099.656565] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1099.656565] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1099.656903] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1099.656903] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1099.656968] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1099.657148] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1099.657300] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1099.657463] env[61936]: DEBUG nova.virt.hardware [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1099.665588] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1099.665944] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-970153a3-6eab-4c24-b2a5-8a0da5cc215d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.687957] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1099.687957] env[61936]: value = "task-1253553" [ 1099.687957] env[61936]: _type = "Task" [ 1099.687957] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.696299] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253553, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.784223] env[61936]: DEBUG nova.network.neutron [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Successfully created port: 5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1099.926912] env[61936]: DEBUG nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1100.090458] env[61936]: DEBUG nova.network.neutron [-] [instance: 484e5d41-29f4-4845-9633-157c03766978] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.120353] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8690c1a7-c369-4229-b372-bd39c3fbcaf5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.130363] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c007e3-c6b5-4779-b96a-06e726ae9ec2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.159263] env[61936]: DEBUG nova.compute.manager [req-b86fac58-2f4d-42a7-9670-a85087041511 req-6dbb4c49-e2fc-4756-a3f8-24fb7ce9d0a8 service nova] [instance: 484e5d41-29f4-4845-9633-157c03766978] Detach interface failed, port_id=7450de29-76d5-40b0-ae76-a79b3455a9bc, reason: Instance 484e5d41-29f4-4845-9633-157c03766978 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1100.198420] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253553, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.271897] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.272383] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.272773] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.273100] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.273380] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.276228] env[61936]: INFO nova.compute.manager [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Terminating instance [ 1100.593757] env[61936]: INFO nova.compute.manager [-] [instance: 484e5d41-29f4-4845-9633-157c03766978] Took 1.44 seconds to deallocate network for instance. [ 1100.699267] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253553, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.780878] env[61936]: DEBUG nova.compute.manager [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1100.781182] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1100.782075] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5b3297-fc49-46bd-a07a-1c9192c223c4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.791512] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.791746] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d3592ef-0150-4fe7-9373-27670dcc4040 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.797615] env[61936]: DEBUG oslo_vmware.api [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1100.797615] env[61936]: value = "task-1253554" [ 1100.797615] env[61936]: _type = "Task" [ 1100.797615] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.805443] env[61936]: DEBUG oslo_vmware.api [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253554, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.937212] env[61936]: DEBUG nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1100.965945] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1100.966220] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.966370] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1100.966556] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.966702] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1100.966849] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1100.967307] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1100.967307] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1100.967589] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1100.967589] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1100.967982] env[61936]: DEBUG nova.virt.hardware [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1100.968657] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c82816-e2b8-46df-9745-e056a8e9073c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.976327] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2897e70b-66a3-48a4-86a7-7d9097ebacd7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.100160] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.100498] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.100765] env[61936]: DEBUG nova.objects.instance [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lazy-loading 'resources' on Instance uuid 484e5d41-29f4-4845-9633-157c03766978 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.201199] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253553, 'name': ReconfigVM_Task, 'duration_secs': 1.157303} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.201442] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.202148] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbeb9b4-bfd3-4932-8463-a3cae1c4702f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.224045] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 6c78d80f-84e8-4250-a60a-c9e28befd07a/6c78d80f-84e8-4250-a60a-c9e28befd07a.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.224310] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecdf2a81-f889-4949-ba13-1cb5c15214d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.242672] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1101.242672] env[61936]: value = "task-1253555" [ 1101.242672] env[61936]: _type = "Task" [ 1101.242672] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.250624] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253555, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.307916] env[61936]: DEBUG oslo_vmware.api [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253554, 'name': PowerOffVM_Task, 'duration_secs': 0.238265} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.308226] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1101.308358] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1101.308667] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44e23d45-d2d0-432f-a60f-a58b06579285 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.317866] env[61936]: DEBUG nova.compute.manager [req-db2f6e99-c0b3-46e3-9a91-fbf508e37768 req-953b7f82-d321-4f9f-b6ca-9fa2f4e9e6f1 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Received event network-vif-plugged-5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1101.318171] env[61936]: DEBUG oslo_concurrency.lockutils [req-db2f6e99-c0b3-46e3-9a91-fbf508e37768 req-953b7f82-d321-4f9f-b6ca-9fa2f4e9e6f1 service nova] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.318306] env[61936]: DEBUG oslo_concurrency.lockutils [req-db2f6e99-c0b3-46e3-9a91-fbf508e37768 req-953b7f82-d321-4f9f-b6ca-9fa2f4e9e6f1 service nova] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.318464] env[61936]: DEBUG oslo_concurrency.lockutils [req-db2f6e99-c0b3-46e3-9a91-fbf508e37768 req-953b7f82-d321-4f9f-b6ca-9fa2f4e9e6f1 service nova] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.318633] env[61936]: DEBUG nova.compute.manager [req-db2f6e99-c0b3-46e3-9a91-fbf508e37768 req-953b7f82-d321-4f9f-b6ca-9fa2f4e9e6f1 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] No waiting events found dispatching network-vif-plugged-5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1101.318801] env[61936]: WARNING nova.compute.manager [req-db2f6e99-c0b3-46e3-9a91-fbf508e37768 req-953b7f82-d321-4f9f-b6ca-9fa2f4e9e6f1 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Received unexpected event network-vif-plugged-5ba0ab28-3350-4b44-9b20-cc64a712f113 for instance with vm_state building and task_state spawning. [ 1101.377366] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1101.377665] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1101.377834] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleting the datastore file [datastore2] fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1101.378116] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29d692b9-da7a-48dd-9c60-55f49ee81c7e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.384338] env[61936]: DEBUG oslo_vmware.api [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1101.384338] env[61936]: value = "task-1253557" [ 1101.384338] env[61936]: _type = "Task" [ 1101.384338] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.391931] env[61936]: DEBUG oslo_vmware.api [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253557, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.462064] env[61936]: DEBUG nova.network.neutron [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Successfully updated port: 5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1101.740795] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ada2bcf-e0f3-4e70-bbfb-a0a90cd98060 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.752413] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253555, 'name': ReconfigVM_Task, 'duration_secs': 0.260045} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.754216] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 6c78d80f-84e8-4250-a60a-c9e28befd07a/6c78d80f-84e8-4250-a60a-c9e28befd07a.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1101.754501] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance '6c78d80f-84e8-4250-a60a-c9e28befd07a' progress to 50 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1101.759264] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557d4576-a231-4a43-a150-cf9f41fbcfd2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.790017] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e934a47-ffcb-42b3-aaac-e6c38bd1dabf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.797461] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43da8dde-b703-48be-ae15-7451d81ed16f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.810697] env[61936]: DEBUG nova.compute.provider_tree [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.894226] env[61936]: DEBUG oslo_vmware.api [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253557, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126453} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.894485] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1101.894671] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1101.894845] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1101.895034] env[61936]: INFO nova.compute.manager [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1101.895274] env[61936]: DEBUG oslo.service.loopingcall [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1101.895463] env[61936]: DEBUG nova.compute.manager [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1101.895560] env[61936]: DEBUG nova.network.neutron [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1101.964728] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.964880] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.965042] env[61936]: DEBUG nova.network.neutron [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.262156] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afeef25-5c00-4132-846b-461403347b27 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.283850] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb0306c-6d20-4f09-a3be-2a1112cb72bb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.304203] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance '6c78d80f-84e8-4250-a60a-c9e28befd07a' progress to 67 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1102.313754] env[61936]: DEBUG nova.scheduler.client.report [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1102.514542] env[61936]: DEBUG nova.network.neutron [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1102.685739] env[61936]: DEBUG nova.network.neutron [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updating instance_info_cache with network_info: [{"id": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "address": "fa:16:3e:6c:df:a3", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba0ab28-33", "ovs_interfaceid": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.818146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.841966] env[61936]: INFO nova.scheduler.client.report [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleted allocations for instance 484e5d41-29f4-4845-9633-157c03766978 [ 1102.843078] env[61936]: DEBUG nova.network.neutron [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.845542] env[61936]: DEBUG nova.network.neutron [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Port c0d12433-0c29-4d8f-9664-dcbb76583dfc binding to destination host cpu-1 is already ACTIVE {{(pid=61936) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1103.188057] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.188418] env[61936]: DEBUG nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Instance network_info: |[{"id": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "address": "fa:16:3e:6c:df:a3", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba0ab28-33", "ovs_interfaceid": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1103.188862] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:df:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ba0ab28-3350-4b44-9b20-cc64a712f113', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.196368] env[61936]: DEBUG oslo.service.loopingcall [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.196578] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.196811] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f04948d-7d29-48b3-97c7-ca603d5c8d11 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.216266] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.216266] env[61936]: value = "task-1253558" [ 1103.216266] env[61936]: _type = "Task" [ 1103.216266] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.223774] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253558, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.348076] env[61936]: DEBUG nova.compute.manager [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Received event network-changed-5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1103.348404] env[61936]: DEBUG nova.compute.manager [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Refreshing instance network info cache due to event network-changed-5ba0ab28-3350-4b44-9b20-cc64a712f113. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1103.348672] env[61936]: DEBUG oslo_concurrency.lockutils [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] Acquiring lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.348857] env[61936]: DEBUG oslo_concurrency.lockutils [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] Acquired lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.349042] env[61936]: DEBUG nova.network.neutron [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Refreshing network info cache for port 5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1103.350269] env[61936]: INFO nova.compute.manager [-] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Took 1.45 seconds to deallocate network for instance. [ 1103.358027] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9835de53-9218-4ebd-95bc-e1bcd590427e tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "484e5d41-29f4-4845-9633-157c03766978" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.825s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.726452] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253558, 'name': CreateVM_Task, 'duration_secs': 0.310679} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.727587] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1103.727587] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.727587] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.727983] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1103.728053] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60a5d129-01bd-430d-a62a-4ee64b4c6699 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.732645] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1103.732645] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526aba97-a7d2-3d85-3f58-aa5edf5da569" [ 1103.732645] env[61936]: _type = "Task" [ 1103.732645] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.739997] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526aba97-a7d2-3d85-3f58-aa5edf5da569, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.868927] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.869307] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.869417] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.872133] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.872343] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.872646] env[61936]: DEBUG nova.objects.instance [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'resources' on Instance uuid fbde880a-b47c-49e8-b84d-0efee21d62c0 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.051295] env[61936]: DEBUG nova.network.neutron [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updated VIF entry in instance network info cache for port 5ba0ab28-3350-4b44-9b20-cc64a712f113. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1104.051718] env[61936]: DEBUG nova.network.neutron [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updating instance_info_cache with network_info: [{"id": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "address": "fa:16:3e:6c:df:a3", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba0ab28-33", "ovs_interfaceid": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.243474] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]526aba97-a7d2-3d85-3f58-aa5edf5da569, 'name': SearchDatastore_Task, 'duration_secs': 0.00977} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.243721] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.243969] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.244229] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.244379] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.244559] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.244828] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01b49100-9da9-43dd-bd81-2410e10546e0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.253359] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.253563] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.254268] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0f8c642-9121-4c14-97a7-66d814af463b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.259693] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1104.259693] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ce45ab-a501-d1ca-cba7-bd870f9e7b4c" [ 1104.259693] env[61936]: _type = "Task" [ 1104.259693] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.267075] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ce45ab-a501-d1ca-cba7-bd870f9e7b4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.508298] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e56cfe-c95f-4f65-93d0-a9041d46a5cc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.515987] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8174d8-4b1e-46cc-8612-925fa21f11d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.545307] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c686bc2f-c1bb-451c-9acb-116e79342f33 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.552841] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d77852-48c6-467c-9037-a7b2c81c2551 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.556666] env[61936]: DEBUG oslo_concurrency.lockutils [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] Releasing lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.556915] env[61936]: DEBUG nova.compute.manager [req-c353a68c-ac28-4c64-aec9-0b1dc3a83f32 req-b727c0d2-3c5c-48a9-b57e-ce9cc8f38c60 service nova] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Received event network-vif-deleted-d169ca50-5a14-4a65-9da0-ea79c80f75c8 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1104.566557] env[61936]: DEBUG nova.compute.provider_tree [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.771699] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52ce45ab-a501-d1ca-cba7-bd870f9e7b4c, 'name': SearchDatastore_Task, 'duration_secs': 0.00927} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.772496] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d06944c3-7c25-4227-a57b-9f65b6eec957 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.777826] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1104.777826] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b9f41-301c-62fd-a328-4701b13c1b28" [ 1104.777826] env[61936]: _type = "Task" [ 1104.777826] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.785038] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b9f41-301c-62fd-a328-4701b13c1b28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.914202] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.914400] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.914582] env[61936]: DEBUG nova.network.neutron [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1105.070272] env[61936]: DEBUG nova.scheduler.client.report [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1105.288115] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529b9f41-301c-62fd-a328-4701b13c1b28, 'name': SearchDatastore_Task, 'duration_secs': 0.009736} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.288406] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.288673] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.288937] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abba1973-3c71-4e80-8089-604ed0187f21 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.295357] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1105.295357] env[61936]: value = "task-1253560" [ 1105.295357] env[61936]: _type = "Task" [ 1105.295357] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.302878] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253560, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.576375] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.600168] env[61936]: INFO nova.scheduler.client.report [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted allocations for instance fbde880a-b47c-49e8-b84d-0efee21d62c0 [ 1105.710956] env[61936]: DEBUG nova.network.neutron [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance_info_cache with network_info: [{"id": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "address": "fa:16:3e:a3:46:a9", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0d12433-0c", "ovs_interfaceid": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.805935] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253560, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454405} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.806226] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1105.806437] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1105.806695] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cea217b6-3250-454b-947a-60a4d49b165c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.812863] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1105.812863] env[61936]: value = "task-1253561" [ 1105.812863] env[61936]: _type = "Task" [ 1105.812863] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.819946] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.109054] env[61936]: DEBUG oslo_concurrency.lockutils [None req-03c219ab-979e-497d-ab72-3942d093bb36 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "fbde880a-b47c-49e8-b84d-0efee21d62c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.837s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.213266] env[61936]: DEBUG oslo_concurrency.lockutils [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.322585] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068398} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.322816] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1106.323598] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcdf68e-3eff-42c2-8eed-32d50779b603 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.345258] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.345488] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fedab059-bc12-4c0a-8b0d-103f9ebee3db {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.364891] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1106.364891] env[61936]: value = "task-1253562" [ 1106.364891] env[61936]: _type = "Task" [ 1106.364891] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.373476] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253562, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.737481] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4594aa-7c5e-457f-8926-660bddfea534 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.758822] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523cb85a-721f-4bbf-a981-0ea224a817be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.766018] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance '6c78d80f-84e8-4250-a60a-c9e28befd07a' progress to 83 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1106.875507] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253562, 'name': ReconfigVM_Task, 'duration_secs': 0.301273} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.877637] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1106.877637] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bd47271-ef23-47a2-92a2-e1be1c1e5bb8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.883238] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1106.883238] env[61936]: value = "task-1253563" [ 1106.883238] env[61936]: _type = "Task" [ 1106.883238] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.891478] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253563, 'name': Rename_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.272474] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1107.272814] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e43d37a0-52d9-4d7f-9951-6ed2d9680e77 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.279303] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1107.279303] env[61936]: value = "task-1253564" [ 1107.279303] env[61936]: _type = "Task" [ 1107.279303] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.286971] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253564, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.392881] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253563, 'name': Rename_Task, 'duration_secs': 0.138125} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.393181] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1107.393446] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecc881bf-f4e5-4b89-99fa-29df601a2798 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.399652] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1107.399652] env[61936]: value = "task-1253565" [ 1107.399652] env[61936]: _type = "Task" [ 1107.399652] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.407969] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253565, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.789073] env[61936]: DEBUG oslo_vmware.api [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253564, 'name': PowerOnVM_Task, 'duration_secs': 0.468469} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.789419] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1107.789581] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-54779024-bd8c-4fca-a0b9-ca28f1d59c29 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance '6c78d80f-84e8-4250-a60a-c9e28befd07a' progress to 100 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1107.910391] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253565, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.410020] env[61936]: DEBUG oslo_vmware.api [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253565, 'name': PowerOnVM_Task, 'duration_secs': 0.556777} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.410827] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.410827] env[61936]: INFO nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Took 7.47 seconds to spawn the instance on the hypervisor. [ 1108.410827] env[61936]: DEBUG nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1108.411512] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f735833d-823c-4632-9cfc-767dca6f5a38 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.617512] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-f1d45dcb-768d-447e-be49-ef43ad98fa61-7e49da52-d495-432c-8a85-36a4301edcd1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.617512] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-f1d45dcb-768d-447e-be49-ef43ad98fa61-7e49da52-d495-432c-8a85-36a4301edcd1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.617832] env[61936]: DEBUG nova.objects.instance [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'flavor' on Instance uuid f1d45dcb-768d-447e-be49-ef43ad98fa61 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.654086] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.654329] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.762370] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.762825] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.763146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.763359] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.763556] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.765827] env[61936]: INFO nova.compute.manager [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Terminating instance [ 1108.929174] env[61936]: INFO nova.compute.manager [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Took 13.52 seconds to build instance. [ 1109.156482] env[61936]: DEBUG nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1109.220025] env[61936]: DEBUG nova.objects.instance [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'pci_requests' on Instance uuid f1d45dcb-768d-447e-be49-ef43ad98fa61 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.269278] env[61936]: DEBUG nova.compute.manager [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1109.269546] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.270443] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60572f5c-abe6-4f17-b808-bd46fac7be74 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.278553] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1109.278783] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d601313-6b35-4fbc-993b-1ee6f1a75477 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.284385] env[61936]: DEBUG oslo_vmware.api [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1109.284385] env[61936]: value = "task-1253566" [ 1109.284385] env[61936]: _type = "Task" [ 1109.284385] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.293154] env[61936]: DEBUG oslo_vmware.api [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.431611] env[61936]: DEBUG oslo_concurrency.lockutils [None req-0ae67886-c9fb-4746-8862-a48fc2173dad tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.030s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.681044] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.681491] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.683056] env[61936]: INFO nova.compute.claims [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1109.685733] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.685945] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.686137] env[61936]: DEBUG nova.compute.manager [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Going to confirm migration 3 {{(pid=61936) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5174}} [ 1109.722164] env[61936]: DEBUG nova.objects.base [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1109.722381] env[61936]: DEBUG nova.network.neutron [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1109.784225] env[61936]: DEBUG nova.policy [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1109.796198] env[61936]: DEBUG oslo_vmware.api [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253566, 'name': PowerOffVM_Task, 'duration_secs': 0.275496} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.796469] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.796660] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.796911] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fe8b945-299f-4884-b7a2-e4643f4ca33b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.859282] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.859538] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.859777] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleting the datastore file [datastore1] f3f2a0a8-cfe3-462f-a940-05580d5da32d {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.860119] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d559d5a1-576c-4d34-b7f8-d0d963508c5b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.866434] env[61936]: DEBUG oslo_vmware.api [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1109.866434] env[61936]: value = "task-1253568" [ 1109.866434] env[61936]: _type = "Task" [ 1109.866434] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.874743] env[61936]: DEBUG oslo_vmware.api [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253568, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.250717] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.251084] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquired lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.251423] env[61936]: DEBUG nova.network.neutron [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1110.251739] env[61936]: DEBUG nova.objects.instance [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'info_cache' on Instance uuid 6c78d80f-84e8-4250-a60a-c9e28befd07a {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.376842] env[61936]: DEBUG oslo_vmware.api [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253568, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228594} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.377180] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1110.377470] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1110.377711] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1110.377896] env[61936]: INFO nova.compute.manager [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1110.378162] env[61936]: DEBUG oslo.service.loopingcall [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.378368] env[61936]: DEBUG nova.compute.manager [-] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1110.378465] env[61936]: DEBUG nova.network.neutron [-] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1110.832342] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77fa805d-c482-492c-a5f9-1aaea61328f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.839794] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca68b140-1680-4a0d-aa65-1052026beac1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.870878] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edfcde5-8a7a-46a2-9199-5ce58f0b3263 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.878788] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4bc758-b08a-4762-a40c-68a9a2369a95 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.892747] env[61936]: DEBUG nova.compute.provider_tree [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.132410] env[61936]: DEBUG nova.network.neutron [-] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.261218] env[61936]: DEBUG nova.network.neutron [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Successfully updated port: 7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1111.396306] env[61936]: DEBUG nova.scheduler.client.report [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1111.452450] env[61936]: DEBUG nova.network.neutron [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance_info_cache with network_info: [{"id": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "address": "fa:16:3e:a3:46:a9", "network": {"id": "3373123e-500f-4c6b-97fd-6c59fa73ec4e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-223907495-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e81b119251964bdb8e5a0ea84b29f2a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0d12433-0c", "ovs_interfaceid": "c0d12433-0c29-4d8f-9664-dcbb76583dfc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.635189] env[61936]: INFO nova.compute.manager [-] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Took 1.26 seconds to deallocate network for instance. [ 1111.765342] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.765639] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.765960] env[61936]: DEBUG nova.network.neutron [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1111.901793] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.220s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.902317] env[61936]: DEBUG nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1111.955321] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Releasing lock "refresh_cache-6c78d80f-84e8-4250-a60a-c9e28befd07a" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.955572] env[61936]: DEBUG nova.objects.instance [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lazy-loading 'migration_context' on Instance uuid 6c78d80f-84e8-4250-a60a-c9e28befd07a {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.141725] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.141998] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.142245] env[61936]: DEBUG nova.objects.instance [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lazy-loading 'resources' on Instance uuid f3f2a0a8-cfe3-462f-a940-05580d5da32d {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.301538] env[61936]: WARNING nova.network.neutron [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] 03166185-d5e0-477b-892e-20bea284c8db already exists in list: networks containing: ['03166185-d5e0-477b-892e-20bea284c8db']. ignoring it [ 1112.406896] env[61936]: DEBUG nova.compute.utils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1112.410466] env[61936]: DEBUG nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1112.410466] env[61936]: DEBUG nova.network.neutron [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1112.456341] env[61936]: DEBUG nova.policy [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e0323c8cb7c4081ac62a7193e209d7b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '905821a1ff2b4011994c0d2d8bc08b13', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1112.458163] env[61936]: DEBUG nova.objects.base [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Object Instance<6c78d80f-84e8-4250-a60a-c9e28befd07a> lazy-loaded attributes: info_cache,migration_context {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1112.460712] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3b3945-e4a9-4eb9-a69e-3a69951fbc80 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.482817] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67e8c799-b8f5-44b8-b7d4-fcd9caccf21e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.488376] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1112.488376] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529dbbb7-784b-2711-47a0-365f40582c61" [ 1112.488376] env[61936]: _type = "Task" [ 1112.488376] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.496536] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529dbbb7-784b-2711-47a0-365f40582c61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.577169] env[61936]: DEBUG nova.network.neutron [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e49da52-d495-432c-8a85-36a4301edcd1", "address": "fa:16:3e:22:8c:da", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e49da52-d4", "ovs_interfaceid": "7e49da52-d495-432c-8a85-36a4301edcd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.730086] env[61936]: DEBUG nova.network.neutron [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Successfully created port: 77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1112.787091] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcf3097-e5b4-462e-be40-8c501a7300b3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.794638] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc73d37-b765-4f84-9150-e1243a177f18 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.824091] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e273198a-d1de-47cd-908a-f20d358ef3be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.831070] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb07e6b2-d13d-437c-b7ee-082c3ece4346 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.843663] env[61936]: DEBUG nova.compute.provider_tree [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.913611] env[61936]: DEBUG nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1112.999321] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529dbbb7-784b-2711-47a0-365f40582c61, 'name': SearchDatastore_Task, 'duration_secs': 0.007227} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.999615] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.080137] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.080782] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.080940] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.081822] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323e0bf2-5388-48a2-83ed-3251a5423efe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.098651] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1113.098882] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.099058] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1113.099280] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.099431] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1113.099577] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1113.099781] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1113.099939] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1113.100129] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1113.100300] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1113.100472] env[61936]: DEBUG nova.virt.hardware [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1113.106742] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Reconfiguring VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1113.107036] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74b46075-0369-445b-b29e-4e5a785a4d66 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.123337] env[61936]: DEBUG oslo_vmware.api [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1113.123337] env[61936]: value = "task-1253569" [ 1113.123337] env[61936]: _type = "Task" [ 1113.123337] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.130512] env[61936]: DEBUG oslo_vmware.api [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253569, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.347031] env[61936]: DEBUG nova.scheduler.client.report [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1113.633629] env[61936]: DEBUG oslo_vmware.api [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253569, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.852537] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.854919] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.855s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.870890] env[61936]: INFO nova.scheduler.client.report [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleted allocations for instance f3f2a0a8-cfe3-462f-a940-05580d5da32d [ 1113.924030] env[61936]: DEBUG nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1113.947701] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1113.947944] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.948114] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1113.948298] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.948445] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1113.948593] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1113.948799] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1113.948954] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1113.949131] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1113.949350] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1113.949557] env[61936]: DEBUG nova.virt.hardware [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1113.950444] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160ea146-1909-4480-9aa5-da7c64aa4ed3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.958512] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caaa6a25-c19e-4535-a6c2-de2fbc78add8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.133663] env[61936]: DEBUG oslo_vmware.api [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253569, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.208089] env[61936]: DEBUG nova.network.neutron [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Successfully updated port: 77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1114.378591] env[61936]: DEBUG oslo_concurrency.lockutils [None req-78d8ee7e-46dd-4e22-893d-52ae3c7f2e4a tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "f3f2a0a8-cfe3-462f-a940-05580d5da32d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.616s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.490389] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58ac2aa-db1b-405a-b155-5e0e23d0b82b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.499770] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3075bad2-cf5c-4f69-bc5c-eca31ff9b0f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.531635] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919e07fa-314b-474b-95cf-50c2d77aa76f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.540009] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731ef74c-037d-4f4e-90c3-0f25ea332980 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.555315] env[61936]: DEBUG nova.compute.provider_tree [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.635313] env[61936]: DEBUG oslo_vmware.api [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253569, 'name': ReconfigVM_Task, 'duration_secs': 1.465533} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.636039] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.636327] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Reconfigured VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1114.710676] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.710850] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.710962] env[61936]: DEBUG nova.network.neutron [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1115.058409] env[61936]: DEBUG nova.scheduler.client.report [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1115.141997] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9d863155-b52d-41d1-b930-df759e06faf3 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-f1d45dcb-768d-447e-be49-ef43ad98fa61-7e49da52-d495-432c-8a85-36a4301edcd1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.524s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.242725] env[61936]: DEBUG nova.network.neutron [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1115.407053] env[61936]: DEBUG nova.network.neutron [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77480d6a-c7", "ovs_interfaceid": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.909986] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.910362] env[61936]: DEBUG nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Instance network_info: |[{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77480d6a-c7", "ovs_interfaceid": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1115.910822] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:49:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77480d6a-c7f5-41bf-9dc1-6554e54c4552', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.918412] env[61936]: DEBUG oslo.service.loopingcall [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.918653] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1115.918883] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1aa261f4-1b7f-4ea7-9c53-305d4320efdb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.940496] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.940496] env[61936]: value = "task-1253570" [ 1115.940496] env[61936]: _type = "Task" [ 1115.940496] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.947953] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253570, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.068553] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.213s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.450447] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253570, 'name': CreateVM_Task} progress is 25%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.482242] env[61936]: DEBUG nova.compute.manager [req-9bd6e560-e0dd-4cb3-84cf-2e8825e967e0 req-2d4c3250-b8ef-47ef-bafd-f125b7cc3ccb service nova] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Received event network-vif-deleted-da841dba-1be9-4d65-949a-5f8c54e14cae {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1116.499576] env[61936]: DEBUG nova.compute.manager [req-cad11714-b51a-4320-a822-cbf375633a50 req-1452faa0-7853-4119-b807-2ac649d52639 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-vif-plugged-7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1116.499859] env[61936]: DEBUG oslo_concurrency.lockutils [req-cad11714-b51a-4320-a822-cbf375633a50 req-1452faa0-7853-4119-b807-2ac649d52639 service nova] Acquiring lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.500068] env[61936]: DEBUG oslo_concurrency.lockutils [req-cad11714-b51a-4320-a822-cbf375633a50 req-1452faa0-7853-4119-b807-2ac649d52639 service nova] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.500261] env[61936]: DEBUG oslo_concurrency.lockutils [req-cad11714-b51a-4320-a822-cbf375633a50 req-1452faa0-7853-4119-b807-2ac649d52639 service nova] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.500328] env[61936]: DEBUG nova.compute.manager [req-cad11714-b51a-4320-a822-cbf375633a50 req-1452faa0-7853-4119-b807-2ac649d52639 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] No waiting events found dispatching network-vif-plugged-7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1116.500487] env[61936]: WARNING nova.compute.manager [req-cad11714-b51a-4320-a822-cbf375633a50 req-1452faa0-7853-4119-b807-2ac649d52639 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received unexpected event network-vif-plugged-7e49da52-d495-432c-8a85-36a4301edcd1 for instance with vm_state active and task_state None. [ 1116.633325] env[61936]: INFO nova.scheduler.client.report [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted allocation for migration 0b3d90bf-182f-478f-bd29-db3c331088fe [ 1116.951950] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253570, 'name': CreateVM_Task, 'duration_secs': 0.659519} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.952142] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1116.952768] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.952932] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.953279] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1116.953543] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65437b57-4a0d-4fb4-a003-aa2e1cfecd46 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.958696] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1116.958696] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a037b7-4820-fbe5-709c-3b16428d5e43" [ 1116.958696] env[61936]: _type = "Task" [ 1116.958696] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.966480] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a037b7-4820-fbe5-709c-3b16428d5e43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.140415] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.454s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.196692] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "fa845f8e-957c-4c0b-a647-190c32989dcd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.196957] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.197195] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "fa845f8e-957c-4c0b-a647-190c32989dcd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.197379] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.197550] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.200335] env[61936]: INFO nova.compute.manager [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Terminating instance [ 1117.257843] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.258130] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.258338] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.258562] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.258692] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.260965] env[61936]: INFO nova.compute.manager [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Terminating instance [ 1117.470037] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a037b7-4820-fbe5-709c-3b16428d5e43, 'name': SearchDatastore_Task, 'duration_secs': 0.009873} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.470037] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.470388] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1117.470388] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.470574] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.470712] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.470973] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-911e5839-4878-48b6-9637-d34554805189 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.479096] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.479300] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1117.479994] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2968b045-e2d3-435b-9b9e-a1ac3abf138d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.484895] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1117.484895] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52af8f7f-b298-e3f9-bf63-3ffc517bf70f" [ 1117.484895] env[61936]: _type = "Task" [ 1117.484895] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.491888] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52af8f7f-b298-e3f9-bf63-3ffc517bf70f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.535051] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-f1d45dcb-768d-447e-be49-ef43ad98fa61-7e49da52-d495-432c-8a85-36a4301edcd1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.535051] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-f1d45dcb-768d-447e-be49-ef43ad98fa61-7e49da52-d495-432c-8a85-36a4301edcd1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.707053] env[61936]: DEBUG nova.compute.manager [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1117.707277] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.708476] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3699c8-4906-4f11-8d18-df286f0ca45c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.717700] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.718470] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b1a2595-e294-4f9f-9907-5eef87c7f3a3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.725879] env[61936]: DEBUG oslo_vmware.api [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1117.725879] env[61936]: value = "task-1253571" [ 1117.725879] env[61936]: _type = "Task" [ 1117.725879] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.734067] env[61936]: DEBUG oslo_vmware.api [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.765658] env[61936]: DEBUG nova.compute.manager [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1117.765658] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.766395] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fae360-3553-4e5f-9a42-17b1c832aeb7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.775670] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.775916] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13e68b4a-25ca-4366-bde8-b9232c3fa267 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.782250] env[61936]: DEBUG oslo_vmware.api [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1117.782250] env[61936]: value = "task-1253572" [ 1117.782250] env[61936]: _type = "Task" [ 1117.782250] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.791430] env[61936]: DEBUG oslo_vmware.api [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.995050] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52af8f7f-b298-e3f9-bf63-3ffc517bf70f, 'name': SearchDatastore_Task, 'duration_secs': 0.007847} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.995319] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-891ff43b-2d99-407d-bbcf-52bf10d46c31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.000884] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1118.000884] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f326d0-f633-1451-b08f-b397acb3cd0e" [ 1118.000884] env[61936]: _type = "Task" [ 1118.000884] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.008482] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f326d0-f633-1451-b08f-b397acb3cd0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.037157] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.037341] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.038126] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7af12da-7bb2-4f5d-b4e3-86bd3fb7900a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.056397] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422925b8-eff2-4496-a1f5-0cba6ff835f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.081574] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Reconfiguring VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1118.081957] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa83dede-0713-4a95-b4a9-c9b3dc656e4d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.100348] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1118.100348] env[61936]: value = "task-1253573" [ 1118.100348] env[61936]: _type = "Task" [ 1118.100348] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.108040] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.207889] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.208197] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.208417] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.208605] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.208776] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.211077] env[61936]: INFO nova.compute.manager [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Terminating instance [ 1118.235180] env[61936]: DEBUG oslo_vmware.api [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253571, 'name': PowerOffVM_Task, 'duration_secs': 0.198386} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.235437] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.235609] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.235853] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce81c5e5-e0a3-4784-9771-fa66e96db726 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.291516] env[61936]: DEBUG oslo_vmware.api [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253572, 'name': PowerOffVM_Task, 'duration_secs': 0.213989} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.291984] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.292126] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.292582] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea77f387-682a-48e4-ba22-b8b7d0fce3c3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.297383] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.297593] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.297771] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleting the datastore file [datastore2] fa845f8e-957c-4c0b-a647-190c32989dcd {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.298019] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e3ccd59-bbae-4988-9125-26407b96a25f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.304180] env[61936]: DEBUG oslo_vmware.api [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for the task: (returnval){ [ 1118.304180] env[61936]: value = "task-1253576" [ 1118.304180] env[61936]: _type = "Task" [ 1118.304180] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.312319] env[61936]: DEBUG oslo_vmware.api [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253576, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.361358] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.361641] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.361819] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Deleting the datastore file [datastore2] 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.362115] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-963fa369-0d41-40cf-b1b4-2364f6026756 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.368167] env[61936]: DEBUG oslo_vmware.api [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1118.368167] env[61936]: value = "task-1253577" [ 1118.368167] env[61936]: _type = "Task" [ 1118.368167] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.375718] env[61936]: DEBUG oslo_vmware.api [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.455463] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.455716] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.455885] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Cleaning up deleted instances {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11626}} [ 1118.511393] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f326d0-f633-1451-b08f-b397acb3cd0e, 'name': SearchDatastore_Task, 'duration_secs': 0.009} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.511694] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.511988] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1118.512314] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6739c352-8d1c-4920-b0bb-caa7a55db936 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.519142] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1118.519142] env[61936]: value = "task-1253578" [ 1118.519142] env[61936]: _type = "Task" [ 1118.519142] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.526774] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253578, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.531955] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-changed-7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1118.532154] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing instance network info cache due to event network-changed-7e49da52-d495-432c-8a85-36a4301edcd1. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1118.532368] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.532570] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.532758] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing network info cache for port 7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1118.613269] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.714993] env[61936]: DEBUG nova.compute.manager [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1118.715304] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1118.716273] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a90f17a-7f8a-447b-8563-29926bb67fe4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.724642] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1118.725076] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74c4a0b4-aa30-465f-833b-9605f6ae3482 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.731708] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1118.731708] env[61936]: value = "task-1253579" [ 1118.731708] env[61936]: _type = "Task" [ 1118.731708] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.740302] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253579, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.815865] env[61936]: DEBUG oslo_vmware.api [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Task: {'id': task-1253576, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153576} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.816221] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1118.816435] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1118.816622] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1118.816830] env[61936]: INFO nova.compute.manager [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1118.817178] env[61936]: DEBUG oslo.service.loopingcall [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1118.817421] env[61936]: DEBUG nova.compute.manager [-] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1118.817472] env[61936]: DEBUG nova.network.neutron [-] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1118.879166] env[61936]: DEBUG oslo_vmware.api [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165202} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.879507] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1118.879708] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1118.879885] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1118.880092] env[61936]: INFO nova.compute.manager [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1118.880406] env[61936]: DEBUG oslo.service.loopingcall [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1118.880627] env[61936]: DEBUG nova.compute.manager [-] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1118.880768] env[61936]: DEBUG nova.network.neutron [-] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1118.981989] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] There are 52 instances to clean {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11635}} [ 1118.982496] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 9f38dcfe-98fd-4dc2-87ce-e44f2c73a6bc] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1119.029866] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253578, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481098} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.030175] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1119.031092] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1119.031516] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b66406f9-3093-42f6-bb3d-7f9d655d4703 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.039679] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1119.039679] env[61936]: value = "task-1253580" [ 1119.039679] env[61936]: _type = "Task" [ 1119.039679] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.054816] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.118012] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.244858] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253579, 'name': PowerOffVM_Task, 'duration_secs': 0.380078} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.245193] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1119.245364] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1119.245624] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-968c496b-8b3f-47fd-abf7-56770a733667 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.304543] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1119.304870] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1119.305155] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleting the datastore file [datastore1] 6c78d80f-84e8-4250-a60a-c9e28befd07a {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1119.305446] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2fdcc5e-b14d-4b56-aca2-55950f74a693 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.309732] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updated VIF entry in instance network info cache for port 7e49da52-d495-432c-8a85-36a4301edcd1. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1119.309912] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e49da52-d495-432c-8a85-36a4301edcd1", "address": "fa:16:3e:22:8c:da", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e49da52-d4", "ovs_interfaceid": "7e49da52-d495-432c-8a85-36a4301edcd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.312349] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for the task: (returnval){ [ 1119.312349] env[61936]: value = "task-1253582" [ 1119.312349] env[61936]: _type = "Task" [ 1119.312349] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.320668] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253582, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.381226] env[61936]: DEBUG nova.compute.manager [req-f1c8eaf3-5c20-4f8f-ad05-d88acd686ea2 req-b46fa3a7-319d-437b-a976-c1ada34226c2 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Received event network-vif-deleted-746d118b-a889-44a8-a6fe-1f8e1c483c33 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1119.381383] env[61936]: INFO nova.compute.manager [req-f1c8eaf3-5c20-4f8f-ad05-d88acd686ea2 req-b46fa3a7-319d-437b-a976-c1ada34226c2 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Neutron deleted interface 746d118b-a889-44a8-a6fe-1f8e1c483c33; detaching it from the instance and deleting it from the info cache [ 1119.381557] env[61936]: DEBUG nova.network.neutron [req-f1c8eaf3-5c20-4f8f-ad05-d88acd686ea2 req-b46fa3a7-319d-437b-a976-c1ada34226c2 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.488433] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: c6043bea-03be-427d-9f39-43f81b0788ea] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1119.551372] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072156} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.551748] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1119.552444] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b243492e-c476-4295-ac4e-fb03cda66891 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.575716] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.575995] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49aed2da-d7e3-4312-8069-381de0747c9e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.596409] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1119.596409] env[61936]: value = "task-1253583" [ 1119.596409] env[61936]: _type = "Task" [ 1119.596409] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.604883] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253583, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.613066] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.687065] env[61936]: DEBUG nova.network.neutron [-] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.813489] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.813755] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received event network-vif-plugged-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1119.813961] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.814195] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.814363] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.814534] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] No waiting events found dispatching network-vif-plugged-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1119.814728] env[61936]: WARNING nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received unexpected event network-vif-plugged-77480d6a-c7f5-41bf-9dc1-6554e54c4552 for instance with vm_state building and task_state spawning. [ 1119.814915] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received event network-changed-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1119.815088] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Refreshing instance network info cache due to event network-changed-77480d6a-c7f5-41bf-9dc1-6554e54c4552. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1119.815280] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Acquiring lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.815420] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Acquired lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.815581] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Refreshing network info cache for port 77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1119.826019] env[61936]: DEBUG oslo_vmware.api [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Task: {'id': task-1253582, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284929} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.826304] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.826493] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.826675] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.826848] env[61936]: INFO nova.compute.manager [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1119.827101] env[61936]: DEBUG oslo.service.loopingcall [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1119.827311] env[61936]: DEBUG nova.compute.manager [-] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1119.827407] env[61936]: DEBUG nova.network.neutron [-] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1119.859120] env[61936]: DEBUG nova.network.neutron [-] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.884456] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b69a0b0-ece4-4807-a8bb-49821ff3225d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.896028] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a64c7f9-63fd-40d2-bb6e-d4ac13ed9db8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.925417] env[61936]: DEBUG nova.compute.manager [req-f1c8eaf3-5c20-4f8f-ad05-d88acd686ea2 req-b46fa3a7-319d-437b-a976-c1ada34226c2 service nova] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Detach interface failed, port_id=746d118b-a889-44a8-a6fe-1f8e1c483c33, reason: Instance 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1119.992456] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: bbb98011-2321-4a69-a882-a4d54df132cc] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1120.106847] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253583, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.115274] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.193181] env[61936]: INFO nova.compute.manager [-] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Took 1.38 seconds to deallocate network for instance. [ 1120.361740] env[61936]: INFO nova.compute.manager [-] [instance: 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8] Took 1.48 seconds to deallocate network for instance. [ 1120.495368] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 160d7915-3795-460d-961f-7b43c8ed0168] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1120.514404] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updated VIF entry in instance network info cache for port 77480d6a-c7f5-41bf-9dc1-6554e54c4552. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1120.514774] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77480d6a-c7", "ovs_interfaceid": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.557330] env[61936]: DEBUG nova.compute.manager [req-cc75726d-1cf7-4f4d-a228-cc051ae330f5 req-3680ee73-0e30-49eb-8c87-841d40c0310f service nova] [instance: fa845f8e-957c-4c0b-a647-190c32989dcd] Received event network-vif-deleted-a6719762-3a94-4d4a-9fea-ff09306d59b3 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1120.583785] env[61936]: DEBUG nova.network.neutron [-] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.607579] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253583, 'name': ReconfigVM_Task, 'duration_secs': 0.685692} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.611222] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.611969] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f62a8751-9253-42e0-8913-60f83cbe200c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.619559] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.621055] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1120.621055] env[61936]: value = "task-1253584" [ 1120.621055] env[61936]: _type = "Task" [ 1120.621055] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.628930] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253584, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.700682] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.701024] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.701292] env[61936]: DEBUG nova.objects.instance [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lazy-loading 'resources' on Instance uuid fa845f8e-957c-4c0b-a647-190c32989dcd {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.868075] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.998798] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: b4fb4cfc-10a5-4b24-9f79-1e6832659f89] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1121.017317] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Releasing lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.017683] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Received event network-changed-5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1121.017865] env[61936]: DEBUG nova.compute.manager [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Refreshing instance network info cache due to event network-changed-5ba0ab28-3350-4b44-9b20-cc64a712f113. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1121.018096] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Acquiring lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.018244] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Acquired lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.018415] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Refreshing network info cache for port 5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1121.086633] env[61936]: INFO nova.compute.manager [-] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Took 1.26 seconds to deallocate network for instance. [ 1121.117112] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.129308] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253584, 'name': Rename_Task, 'duration_secs': 0.13404} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.129571] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1121.129809] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46bafa9b-dbc6-4127-9d04-d856997985bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.135291] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1121.135291] env[61936]: value = "task-1253585" [ 1121.135291] env[61936]: _type = "Task" [ 1121.135291] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.142623] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.318448] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea70a84f-595a-4a07-b39a-afc9cdf7fa14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.326209] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357ade39-c88a-47b4-b6eb-a8be8897a145 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.356655] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdafbae7-76e1-44d1-a185-ded2b3dbe145 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.364190] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d74ce2c-893c-4ac0-bb97-5d2221079a41 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.377761] env[61936]: DEBUG nova.compute.provider_tree [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.408652] env[61936]: DEBUG nova.compute.manager [req-fed74351-6f95-41fa-88c9-d807db85e3be req-8271beb5-e9a2-4a98-b2cf-337d0f6793e4 service nova] [instance: 6c78d80f-84e8-4250-a60a-c9e28befd07a] Received event network-vif-deleted-c0d12433-0c29-4d8f-9664-dcbb76583dfc {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1121.503048] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 53605e96-a4d1-4a7b-8fef-1f1878b099c3] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1121.593789] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.618205] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.643916] env[61936]: DEBUG oslo_vmware.api [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253585, 'name': PowerOnVM_Task, 'duration_secs': 0.425592} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.644994] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.645252] env[61936]: INFO nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1121.645443] env[61936]: DEBUG nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1121.646220] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524e887d-f1d9-41cf-a26b-a6d502e3db8f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.742962] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updated VIF entry in instance network info cache for port 5ba0ab28-3350-4b44-9b20-cc64a712f113. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1121.743342] env[61936]: DEBUG nova.network.neutron [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updating instance_info_cache with network_info: [{"id": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "address": "fa:16:3e:6c:df:a3", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ba0ab28-33", "ovs_interfaceid": "5ba0ab28-3350-4b44-9b20-cc64a712f113", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.881295] env[61936]: DEBUG nova.scheduler.client.report [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1122.006151] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 1c391c45-a041-422d-9a44-a29306f99a6d] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1122.120156] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.164674] env[61936]: INFO nova.compute.manager [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Took 12.50 seconds to build instance. [ 1122.246516] env[61936]: DEBUG oslo_concurrency.lockutils [req-ea3a76fb-a1e4-4aab-95d0-141d104f377b req-e1d8451d-5c54-4867-89e5-a83d56f82bc4 service nova] Releasing lock "refresh_cache-3e337ad6-194b-4cd5-bfee-708ac00b7430" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.386867] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.389776] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.521s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.389776] env[61936]: DEBUG nova.objects.instance [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'resources' on Instance uuid 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.405230] env[61936]: INFO nova.scheduler.client.report [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Deleted allocations for instance fa845f8e-957c-4c0b-a647-190c32989dcd [ 1122.510106] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 18399aa8-b4e4-4b2b-bb3d-b18c18b2d50b] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1122.582589] env[61936]: DEBUG nova.compute.manager [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received event network-changed-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1122.582665] env[61936]: DEBUG nova.compute.manager [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Refreshing instance network info cache due to event network-changed-77480d6a-c7f5-41bf-9dc1-6554e54c4552. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1122.582830] env[61936]: DEBUG oslo_concurrency.lockutils [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] Acquiring lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.582977] env[61936]: DEBUG oslo_concurrency.lockutils [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] Acquired lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.583153] env[61936]: DEBUG nova.network.neutron [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Refreshing network info cache for port 77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1122.621220] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.667510] env[61936]: DEBUG oslo_concurrency.lockutils [None req-699c7a61-168f-4e16-a2f5-f24cd0cd386e tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.013s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.915043] env[61936]: DEBUG oslo_concurrency.lockutils [None req-73ef5244-c969-4038-b6ab-b9ea8d3e01b6 tempest-ServerRescueNegativeTestJSON-178824487 tempest-ServerRescueNegativeTestJSON-178824487-project-member] Lock "fa845f8e-957c-4c0b-a647-190c32989dcd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.718s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.001572] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa751e12-f199-43df-bb48-387f466b7778 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.009961] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b19c96c-fa38-4103-93e9-8a8e3ebbd670 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.013246] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 755c5fea-e3d2-4cb3-b717-5f862401d7e1] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1123.052057] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff238404-c0d4-4fe7-a847-5dc3603d72f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.060161] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73228bd0-5572-4521-bee5-583bc8037779 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.075625] env[61936]: DEBUG nova.compute.provider_tree [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.122335] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.517360] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 8b6696e5-ef1b-4825-903f-de4dfb9bf2a6] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1123.532368] env[61936]: DEBUG nova.network.neutron [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updated VIF entry in instance network info cache for port 77480d6a-c7f5-41bf-9dc1-6554e54c4552. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1123.532765] env[61936]: DEBUG nova.network.neutron [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77480d6a-c7", "ovs_interfaceid": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.578841] env[61936]: DEBUG nova.scheduler.client.report [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1123.622715] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.021439] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 90c2fd2c-937c-4b92-8bb1-6387652a03bb] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1124.036111] env[61936]: DEBUG oslo_concurrency.lockutils [req-177818db-55cc-45a4-a889-c1825675a68e req-769be89c-65e8-469b-9df8-9dbde19da3ca service nova] Releasing lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.084363] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.088296] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.495s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.089038] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.118596] env[61936]: INFO nova.scheduler.client.report [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Deleted allocations for instance 0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8 [ 1124.122062] env[61936]: INFO nova.scheduler.client.report [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Deleted allocations for instance 6c78d80f-84e8-4250-a60a-c9e28befd07a [ 1124.133927] env[61936]: DEBUG oslo_vmware.api [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253573, 'name': ReconfigVM_Task, 'duration_secs': 5.84039} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.135741] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.135741] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Reconfigured VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1124.528854] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 484e5d41-29f4-4845-9633-157c03766978] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1124.629854] env[61936]: DEBUG oslo_concurrency.lockutils [None req-dd59a93d-c582-4dc9-8f70-9b5d0931e141 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "0b5f1180-f1d2-4f7c-8658-64fab0d4ffd8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.371s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.637350] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e040fadb-200d-48a6-937a-5157ed32e003 tempest-DeleteServersTestJSON-277657093 tempest-DeleteServersTestJSON-277657093-project-member] Lock "6c78d80f-84e8-4250-a60a-c9e28befd07a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.429s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.036183] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 62b33dd9-e523-4309-94e6-59671fcfbaea] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1125.540064] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 0137305a-dace-4eda-9d90-7233c001176f] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1125.574429] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.574631] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.574816] env[61936]: DEBUG nova.network.neutron [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.045008] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 8d6c05f2-4bac-41ce-a9f0-4df3d86eba16] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1126.280646] env[61936]: INFO nova.network.neutron [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Port 7e49da52-d495-432c-8a85-36a4301edcd1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1126.281214] env[61936]: DEBUG nova.network.neutron [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.546909] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: fbde880a-b47c-49e8-b84d-0efee21d62c0] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1126.761412] env[61936]: DEBUG nova.compute.manager [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1126.761412] env[61936]: DEBUG nova.compute.manager [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing instance network info cache due to event network-changed-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1126.761412] env[61936]: DEBUG oslo_concurrency.lockutils [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] Acquiring lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.789969] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.789969] env[61936]: DEBUG oslo_concurrency.lockutils [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] Acquired lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.790113] env[61936]: DEBUG nova.network.neutron [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Refreshing network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.050330] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: a99a6fef-3cef-409a-b001-4aca97f852c0] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1127.163800] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.164071] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.292463] env[61936]: DEBUG oslo_concurrency.lockutils [None req-9bd13780-9b3e-4c02-bb3d-2da1ae90bc65 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-f1d45dcb-768d-447e-be49-ef43ad98fa61-7e49da52-d495-432c-8a85-36a4301edcd1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.758s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.320637] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.321127] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.417085] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-7e49da52-d495-432c-8a85-36a4301edcd1" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.417307] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-7e49da52-d495-432c-8a85-36a4301edcd1" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.417645] env[61936]: DEBUG nova.objects.instance [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'flavor' on Instance uuid ea74d4a3-556f-4290-93ea-7f55f8c8bc2c {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.477341] env[61936]: DEBUG nova.network.neutron [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updated VIF entry in instance network info cache for port 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1127.477702] env[61936]: DEBUG nova.network.neutron [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [{"id": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "address": "fa:16:3e:d4:b9:09", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e9fc5fb-74", "ovs_interfaceid": "4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.560308] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 45fd1d4d-7125-43a7-afe7-53f737fb9f8e] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1127.667641] env[61936]: DEBUG nova.compute.utils [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.823066] env[61936]: DEBUG nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1127.982238] env[61936]: DEBUG oslo_concurrency.lockutils [req-bb25f4b8-189f-44e5-82b2-27d53277164c req-f849cdc6-db10-4c5f-bb86-dc4202b90975 service nova] Releasing lock "refresh_cache-f1d45dcb-768d-447e-be49-ef43ad98fa61" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.012160] env[61936]: DEBUG nova.objects.instance [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'pci_requests' on Instance uuid ea74d4a3-556f-4290-93ea-7f55f8c8bc2c {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.063863] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 1d25e242-3542-4707-9112-3711e17df577] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1128.170624] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.445767] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.446061] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.447612] env[61936]: INFO nova.compute.claims [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1128.514823] env[61936]: DEBUG nova.objects.base [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1128.514992] env[61936]: DEBUG nova.network.neutron [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1128.566842] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: b780470b-57d0-43e5-be52-539c78cada6b] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1128.578537] env[61936]: DEBUG nova.policy [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf152d1c3b9242bf94fbbba2ff9ed0c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b9b0b8e03e6540a48947eebb51ad1135', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1128.830692] env[61936]: DEBUG nova.compute.manager [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1128.830946] env[61936]: DEBUG nova.compute.manager [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing instance network info cache due to event network-changed-ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1128.831054] env[61936]: DEBUG oslo_concurrency.lockutils [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.831204] env[61936]: DEBUG oslo_concurrency.lockutils [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.831366] env[61936]: DEBUG nova.network.neutron [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1129.070249] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: f3f2a0a8-cfe3-462f-a940-05580d5da32d] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1129.233409] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.233879] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.234309] env[61936]: INFO nova.compute.manager [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Attaching volume e3022471-aac4-4e76-b26d-a39946b28b03 to /dev/sdb [ 1129.271267] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fe9e08-8a72-4ab3-9b64-9aceb22d6a09 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.278847] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a248cb8e-376d-4bbe-8d46-4eb96510bfd6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.292880] env[61936]: DEBUG nova.virt.block_device [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updating existing volume attachment record: bc838384-a6b0-4c9b-a123-66e9af826456 {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1129.570828] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0e152a-b8b4-4f4f-8c7a-8cf304287dc2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.575196] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 9bd97f84-7c21-44ad-9ed0-d7e4097e264e] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1129.583327] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7dad1d3-d4b3-4116-9701-df0e42e5abee {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.615243] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12812488-22ff-4f1a-ba33-a9319666e672 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.623414] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad4ea31-de12-44d9-adfa-84e9d8c815da {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.637940] env[61936]: DEBUG nova.compute.provider_tree [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.705112] env[61936]: DEBUG nova.network.neutron [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updated VIF entry in instance network info cache for port ef5dde7f-60d3-40b7-877c-53c75f17e14e. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1129.705497] env[61936]: DEBUG nova.network.neutron [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.009589] env[61936]: DEBUG nova.compute.manager [req-cb55e1bf-409a-4867-a763-9b9ec9dd6b9d req-051d2fc4-bc8c-434a-991f-b4683d908ab9 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-vif-plugged-7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1130.009870] env[61936]: DEBUG oslo_concurrency.lockutils [req-cb55e1bf-409a-4867-a763-9b9ec9dd6b9d req-051d2fc4-bc8c-434a-991f-b4683d908ab9 service nova] Acquiring lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.010176] env[61936]: DEBUG oslo_concurrency.lockutils [req-cb55e1bf-409a-4867-a763-9b9ec9dd6b9d req-051d2fc4-bc8c-434a-991f-b4683d908ab9 service nova] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.010366] env[61936]: DEBUG oslo_concurrency.lockutils [req-cb55e1bf-409a-4867-a763-9b9ec9dd6b9d req-051d2fc4-bc8c-434a-991f-b4683d908ab9 service nova] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.010537] env[61936]: DEBUG nova.compute.manager [req-cb55e1bf-409a-4867-a763-9b9ec9dd6b9d req-051d2fc4-bc8c-434a-991f-b4683d908ab9 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] No waiting events found dispatching network-vif-plugged-7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1130.010708] env[61936]: WARNING nova.compute.manager [req-cb55e1bf-409a-4867-a763-9b9ec9dd6b9d req-051d2fc4-bc8c-434a-991f-b4683d908ab9 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received unexpected event network-vif-plugged-7e49da52-d495-432c-8a85-36a4301edcd1 for instance with vm_state active and task_state None. [ 1130.078267] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 1a12fd59-bacd-42b3-99c1-8910f349e2cb] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1130.108019] env[61936]: DEBUG nova.network.neutron [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Successfully updated port: 7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1130.142612] env[61936]: DEBUG nova.scheduler.client.report [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1130.208260] env[61936]: DEBUG oslo_concurrency.lockutils [req-003973d2-196f-4809-9d01-2b0061baea5e req-f15918ac-4009-43d3-a465-d4264766a082 service nova] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.585620] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 2797bce9-f221-49b5-ab2b-42df3f347497] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1130.611862] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.612104] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.612361] env[61936]: DEBUG nova.network.neutron [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1130.652124] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.653618] env[61936]: DEBUG nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1131.089367] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 10669072-a8e8-4262-b390-8bf31af6fafe] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1131.146767] env[61936]: WARNING nova.network.neutron [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] 03166185-d5e0-477b-892e-20bea284c8db already exists in list: networks containing: ['03166185-d5e0-477b-892e-20bea284c8db']. ignoring it [ 1131.158070] env[61936]: DEBUG nova.compute.utils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1131.159308] env[61936]: DEBUG nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1131.159474] env[61936]: DEBUG nova.network.neutron [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1131.205954] env[61936]: DEBUG nova.policy [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cbdbed93b974583bae04b87b39e49a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9ca963bd69594788806a9900942f895b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1131.414074] env[61936]: DEBUG nova.network.neutron [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e49da52-d495-432c-8a85-36a4301edcd1", "address": "fa:16:3e:22:8c:da", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e49da52-d4", "ovs_interfaceid": "7e49da52-d495-432c-8a85-36a4301edcd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.469414] env[61936]: DEBUG nova.network.neutron [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Successfully created port: 2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1131.592415] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 9f6a1c42-e5a5-4f1d-9740-35dc6028183a] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1131.664019] env[61936]: DEBUG nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1131.917205] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.917949] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.918058] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.918916] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9b2e42-cc7f-4337-890c-36371d13729a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.936571] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1131.936779] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1131.936965] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1131.937371] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1131.937371] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1131.937447] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1131.937638] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1131.937800] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1131.937967] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1131.938143] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1131.938323] env[61936]: DEBUG nova.virt.hardware [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1131.946220] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Reconfiguring VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1131.946220] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3d36af7-9b5f-4d12-88c5-e6bd6ab5b96b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.962921] env[61936]: DEBUG oslo_vmware.api [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1131.962921] env[61936]: value = "task-1253589" [ 1131.962921] env[61936]: _type = "Task" [ 1131.962921] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.969398] env[61936]: DEBUG oslo_vmware.api [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253589, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.085049] env[61936]: DEBUG nova.compute.manager [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-changed-7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1132.085248] env[61936]: DEBUG nova.compute.manager [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing instance network info cache due to event network-changed-7e49da52-d495-432c-8a85-36a4301edcd1. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1132.085442] env[61936]: DEBUG oslo_concurrency.lockutils [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.085589] env[61936]: DEBUG oslo_concurrency.lockutils [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.085750] env[61936]: DEBUG nova.network.neutron [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Refreshing network info cache for port 7e49da52-d495-432c-8a85-36a4301edcd1 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1132.096121] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 628882ff-2f0f-49ae-8454-d3791fb55545] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1132.473770] env[61936]: DEBUG oslo_vmware.api [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253589, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.599388] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: f71581c9-b754-4ee4-bf47-ea8e52e4a2f6] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1132.673618] env[61936]: DEBUG nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1132.699379] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1132.699748] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1132.699943] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1132.700200] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1132.700394] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1132.700568] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1132.700829] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1132.701022] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1132.701275] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1132.701443] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1132.701703] env[61936]: DEBUG nova.virt.hardware [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1132.702874] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c34109e-d72e-43fe-8527-f99e01350909 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.711265] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbe4f6c-d783-4a56-9f8a-80367fefc622 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.791189] env[61936]: DEBUG nova.network.neutron [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updated VIF entry in instance network info cache for port 7e49da52-d495-432c-8a85-36a4301edcd1. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1132.791694] env[61936]: DEBUG nova.network.neutron [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7e49da52-d495-432c-8a85-36a4301edcd1", "address": "fa:16:3e:22:8c:da", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e49da52-d4", "ovs_interfaceid": "7e49da52-d495-432c-8a85-36a4301edcd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.853793] env[61936]: DEBUG nova.compute.manager [req-d04b1ad7-fb34-4fcc-a57e-5b5ff5e51844 req-c097d5ff-bf62-4f92-9ec3-008a3b8213e7 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Received event network-vif-plugged-2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1132.854047] env[61936]: DEBUG oslo_concurrency.lockutils [req-d04b1ad7-fb34-4fcc-a57e-5b5ff5e51844 req-c097d5ff-bf62-4f92-9ec3-008a3b8213e7 service nova] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.854262] env[61936]: DEBUG oslo_concurrency.lockutils [req-d04b1ad7-fb34-4fcc-a57e-5b5ff5e51844 req-c097d5ff-bf62-4f92-9ec3-008a3b8213e7 service nova] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.854428] env[61936]: DEBUG oslo_concurrency.lockutils [req-d04b1ad7-fb34-4fcc-a57e-5b5ff5e51844 req-c097d5ff-bf62-4f92-9ec3-008a3b8213e7 service nova] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.854595] env[61936]: DEBUG nova.compute.manager [req-d04b1ad7-fb34-4fcc-a57e-5b5ff5e51844 req-c097d5ff-bf62-4f92-9ec3-008a3b8213e7 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] No waiting events found dispatching network-vif-plugged-2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1132.854759] env[61936]: WARNING nova.compute.manager [req-d04b1ad7-fb34-4fcc-a57e-5b5ff5e51844 req-c097d5ff-bf62-4f92-9ec3-008a3b8213e7 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Received unexpected event network-vif-plugged-2d477dba-b68d-487f-89a8-216750d10485 for instance with vm_state building and task_state spawning. [ 1132.972017] env[61936]: DEBUG oslo_vmware.api [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253589, 'name': ReconfigVM_Task, 'duration_secs': 0.6099} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.972524] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.972768] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Reconfigured VM to attach interface {{(pid=61936) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1133.102516] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 3af90345-f5bf-4d25-90ac-c5389aef5c8e] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1133.294052] env[61936]: DEBUG oslo_concurrency.lockutils [req-e2fd17db-24dd-45f6-a4a1-3b89494a48ef req-26441fc5-116e-419e-813e-8d608ee1fb93 service nova] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.431332] env[61936]: DEBUG nova.network.neutron [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Successfully updated port: 2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1133.456224] env[61936]: DEBUG nova.compute.manager [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Received event network-changed-2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1133.456439] env[61936]: DEBUG nova.compute.manager [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Refreshing instance network info cache due to event network-changed-2d477dba-b68d-487f-89a8-216750d10485. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1133.456713] env[61936]: DEBUG oslo_concurrency.lockutils [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] Acquiring lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.456889] env[61936]: DEBUG oslo_concurrency.lockutils [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] Acquired lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.457107] env[61936]: DEBUG nova.network.neutron [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Refreshing network info cache for port 2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1133.477573] env[61936]: DEBUG oslo_concurrency.lockutils [None req-546eaffb-9223-4c37-8b3d-d2f70533be8b tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-7e49da52-d495-432c-8a85-36a4301edcd1" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.060s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.606132] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: a8b831c7-bd6a-4218-b19e-1c43e3678c59] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1133.837440] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1133.837687] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270045', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'name': 'volume-e3022471-aac4-4e76-b26d-a39946b28b03', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8e0c45d-341e-4ad4-8941-4532ebc53e40', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'serial': 'e3022471-aac4-4e76-b26d-a39946b28b03'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1133.838587] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd81b1c2-6ddf-49f9-a62f-48cbf2321696 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.856274] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab010ca-5c06-4b25-8ad3-928b6acc424b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.880351] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-e3022471-aac4-4e76-b26d-a39946b28b03/volume-e3022471-aac4-4e76-b26d-a39946b28b03.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.880634] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f22393f-1123-4b43-8e71-fab72a5ed8da {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.898272] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1133.898272] env[61936]: value = "task-1253590" [ 1133.898272] env[61936]: _type = "Task" [ 1133.898272] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.905709] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.933612] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.989583] env[61936]: DEBUG nova.network.neutron [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1134.068402] env[61936]: DEBUG nova.network.neutron [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.109196] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 3b09b9f5-9084-4bf4-a441-fb4b2d13c73e] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1134.409423] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253590, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.570987] env[61936]: DEBUG oslo_concurrency.lockutils [req-97a5ddfb-7d64-4d35-85f8-14b54e0941e6 req-0e72b77a-ff46-446f-9b96-adf320466702 service nova] Releasing lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.571395] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.571559] env[61936]: DEBUG nova.network.neutron [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1134.612398] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: efdce04f-b0aa-49a6-94b8-1c960ac33fea] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1134.909918] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253590, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.098021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "interface-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-7e49da52-d495-432c-8a85-36a4301edcd1" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.098021] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-7e49da52-d495-432c-8a85-36a4301edcd1" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.101589] env[61936]: DEBUG nova.network.neutron [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1135.116044] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: aa37292e-8499-46aa-8021-cb8d88b5a35a] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1135.227760] env[61936]: DEBUG nova.network.neutron [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updating instance_info_cache with network_info: [{"id": "2d477dba-b68d-487f-89a8-216750d10485", "address": "fa:16:3e:7d:5b:4b", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d477dba-b6", "ovs_interfaceid": "2d477dba-b68d-487f-89a8-216750d10485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.409995] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253590, 'name': ReconfigVM_Task, 'duration_secs': 1.361233} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.410309] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-e3022471-aac4-4e76-b26d-a39946b28b03/volume-e3022471-aac4-4e76-b26d-a39946b28b03.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.415187] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71d861b1-d9dc-4add-9ff0-228e478fa90f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.430385] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1135.430385] env[61936]: value = "task-1253591" [ 1135.430385] env[61936]: _type = "Task" [ 1135.430385] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.439175] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253591, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.601107] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.601345] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.602327] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4514ace9-563c-436c-912a-9c54c4c56554 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.619759] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 258fed0b-ba19-4c90-981a-96bb36ba8bc8] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1135.622198] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d221467-0d4c-437f-8c8a-259c3f7b6a12 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.648300] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Reconfiguring VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1135.649127] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec076fe1-2a85-4091-afe6-c169568983d0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.666975] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1135.666975] env[61936]: value = "task-1253592" [ 1135.666975] env[61936]: _type = "Task" [ 1135.666975] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.676277] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.731227] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.731565] env[61936]: DEBUG nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Instance network_info: |[{"id": "2d477dba-b68d-487f-89a8-216750d10485", "address": "fa:16:3e:7d:5b:4b", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d477dba-b6", "ovs_interfaceid": "2d477dba-b68d-487f-89a8-216750d10485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1135.732061] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:5b:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd72ef32-a57c-43b0-93df-e8a030987d44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d477dba-b68d-487f-89a8-216750d10485', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1135.739483] env[61936]: DEBUG oslo.service.loopingcall [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1135.739713] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1135.739953] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1af22bc6-0701-47c1-99e9-6be31adbd2be {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.759739] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1135.759739] env[61936]: value = "task-1253593" [ 1135.759739] env[61936]: _type = "Task" [ 1135.759739] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.766844] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253593, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.940925] env[61936]: DEBUG oslo_vmware.api [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253591, 'name': ReconfigVM_Task, 'duration_secs': 0.147413} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.941343] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270045', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'name': 'volume-e3022471-aac4-4e76-b26d-a39946b28b03', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8e0c45d-341e-4ad4-8941-4532ebc53e40', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'serial': 'e3022471-aac4-4e76-b26d-a39946b28b03'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1136.126330] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: c4f17d38-25a3-4630-8bd4-bb40fa6fddee] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1136.176905] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.269822] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253593, 'name': CreateVM_Task, 'duration_secs': 0.278057} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.270022] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1136.270628] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.270797] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.271137] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1136.271386] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f0b1967-c8bb-4581-9d86-3157f6e0c148 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.275550] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1136.275550] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52080e12-0cb4-441b-65fe-b0546486516e" [ 1136.275550] env[61936]: _type = "Task" [ 1136.275550] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.284766] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52080e12-0cb4-441b-65fe-b0546486516e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.629840] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 18175943-35d2-4792-a8ea-e54b4ffbaf2f] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1136.678069] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.785093] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52080e12-0cb4-441b-65fe-b0546486516e, 'name': SearchDatastore_Task, 'duration_secs': 0.008833} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.785650] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.785895] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1136.786142] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.786283] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.786458] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1136.786709] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dffdb5ab-dbdd-4439-a68c-1bed872b2c42 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.794131] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1136.794313] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1136.794972] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96f10b90-0bec-4d07-8a9f-4a3044ff9807 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.799371] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1136.799371] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5279040c-9f75-e878-8991-63a3c85dc945" [ 1136.799371] env[61936]: _type = "Task" [ 1136.799371] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.806147] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5279040c-9f75-e878-8991-63a3c85dc945, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.975935] env[61936]: DEBUG nova.objects.instance [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'flavor' on Instance uuid b8e0c45d-341e-4ad4-8941-4532ebc53e40 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.133239] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: e15be950-1c32-4801-b4e4-a4d31ff0cfec] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1137.179369] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.309951] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5279040c-9f75-e878-8991-63a3c85dc945, 'name': SearchDatastore_Task, 'duration_secs': 0.007819} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.310736] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99f5668d-b878-4efd-9c79-c73dde78dbc4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.315689] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1137.315689] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52db0adb-4606-6f9a-b1c6-5ee6ba495e56" [ 1137.315689] env[61936]: _type = "Task" [ 1137.315689] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.323299] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52db0adb-4606-6f9a-b1c6-5ee6ba495e56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.480441] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2a8bd159-1609-4d83-b609-cb517cacdc3b tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.247s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.637413] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: e800f5e9-deb0-4592-aded-3768d238e311] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1137.679195] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.691719] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.691991] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.826906] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52db0adb-4606-6f9a-b1c6-5ee6ba495e56, 'name': SearchDatastore_Task, 'duration_secs': 0.013088} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.827151] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.827383] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496/2c20a6cb-8d9b-4d4a-9c5e-3a3774292496.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1137.827647] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8adef52-b877-4a54-af2c-32adf846a7c8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.834342] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1137.834342] env[61936]: value = "task-1253594" [ 1137.834342] env[61936]: _type = "Task" [ 1137.834342] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.841368] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253594, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.140601] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: f2844173-4114-4158-99ad-ac651efbb607] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1138.187861] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.194943] env[61936]: INFO nova.compute.manager [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Detaching volume e3022471-aac4-4e76-b26d-a39946b28b03 [ 1138.230252] env[61936]: INFO nova.virt.block_device [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Attempting to driver detach volume e3022471-aac4-4e76-b26d-a39946b28b03 from mountpoint /dev/sdb [ 1138.230541] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1138.230810] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270045', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'name': 'volume-e3022471-aac4-4e76-b26d-a39946b28b03', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8e0c45d-341e-4ad4-8941-4532ebc53e40', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'serial': 'e3022471-aac4-4e76-b26d-a39946b28b03'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1138.231960] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193ebe63-53bb-44e2-ac79-df710a792f47 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.254997] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f403910f-5330-4389-93e8-16587d2a0125 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.262132] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ac0a0e-0c81-4ebb-8e74-404114e8786a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.281822] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9e9ff1-3981-47fc-bb2a-b4da702ce79f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.295643] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] The volume has not been displaced from its original location: [datastore1] volume-e3022471-aac4-4e76-b26d-a39946b28b03/volume-e3022471-aac4-4e76-b26d-a39946b28b03.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1138.300776] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1138.301050] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21c476ba-a3d5-4c08-ad43-84306075c9ba {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.318276] env[61936]: DEBUG oslo_vmware.api [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1138.318276] env[61936]: value = "task-1253595" [ 1138.318276] env[61936]: _type = "Task" [ 1138.318276] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.325890] env[61936]: DEBUG oslo_vmware.api [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253595, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.342662] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253594, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.421446} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.342923] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore2] 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496/2c20a6cb-8d9b-4d4a-9c5e-3a3774292496.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1138.343200] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1138.343482] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f91b59bf-a488-44b6-a022-4b685a857cf9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.349037] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1138.349037] env[61936]: value = "task-1253596" [ 1138.349037] env[61936]: _type = "Task" [ 1138.349037] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.358752] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253596, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.646388] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: e5e3e7af-8104-4f7a-ae18-81e03a932608] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1138.680377] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.827713] env[61936]: DEBUG oslo_vmware.api [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253595, 'name': ReconfigVM_Task, 'duration_secs': 0.207468} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.828038] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1138.832620] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90e249ba-0c32-4285-bb6f-0831cc63de12 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.847045] env[61936]: DEBUG oslo_vmware.api [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1138.847045] env[61936]: value = "task-1253597" [ 1138.847045] env[61936]: _type = "Task" [ 1138.847045] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.859648] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253596, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.053914} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.859873] env[61936]: DEBUG oslo_vmware.api [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253597, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.860136] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1138.860824] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecace2e0-b738-43b5-a0f2-60a03df29c78 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.881454] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496/2c20a6cb-8d9b-4d4a-9c5e-3a3774292496.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1138.881710] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e84dcce7-f592-46d8-9a20-72e6f71530d3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.900614] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1138.900614] env[61936]: value = "task-1253598" [ 1138.900614] env[61936]: _type = "Task" [ 1138.900614] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.907895] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253598, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.151560] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 644ccb42-44da-43e8-8b3b-9bfd2ac19a5d] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1139.180844] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.356449] env[61936]: DEBUG oslo_vmware.api [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253597, 'name': ReconfigVM_Task, 'duration_secs': 0.131619} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.356755] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270045', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'name': 'volume-e3022471-aac4-4e76-b26d-a39946b28b03', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8e0c45d-341e-4ad4-8941-4532ebc53e40', 'attached_at': '', 'detached_at': '', 'volume_id': 'e3022471-aac4-4e76-b26d-a39946b28b03', 'serial': 'e3022471-aac4-4e76-b26d-a39946b28b03'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1139.409845] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253598, 'name': ReconfigVM_Task, 'duration_secs': 0.297583} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.410115] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496/2c20a6cb-8d9b-4d4a-9c5e-3a3774292496.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1139.410765] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-806dc231-df42-43a8-9ba6-aed377dd61e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.416911] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1139.416911] env[61936]: value = "task-1253599" [ 1139.416911] env[61936]: _type = "Task" [ 1139.416911] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.424419] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253599, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.654668] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: fafc995e-4c93-4ca0-b078-24a1ae0ab427] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1139.683464] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.909296] env[61936]: DEBUG nova.objects.instance [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'flavor' on Instance uuid b8e0c45d-341e-4ad4-8941-4532ebc53e40 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.929295] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253599, 'name': Rename_Task, 'duration_secs': 0.158173} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.930414] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1139.930678] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a77139b7-da3f-4512-9681-15c32da2330e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.937494] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1139.937494] env[61936]: value = "task-1253600" [ 1139.937494] env[61936]: _type = "Task" [ 1139.937494] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.945597] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253600, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.157709] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 77a59650-d28b-4862-bb35-6910e6f7d6c9] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1140.182720] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.448123] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253600, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.661591] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: bdea3742-5789-4e97-b139-33cfee134843] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1140.683330] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.916693] env[61936]: DEBUG oslo_concurrency.lockutils [None req-14abfa4a-44e3-4b80-a919-0feff7f236b9 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.225s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.949088] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253600, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.165223] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 5526cbd5-b1ad-453b-8401-eee7aa356606] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1141.184892] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.453431] env[61936]: DEBUG oslo_vmware.api [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253600, 'name': PowerOnVM_Task, 'duration_secs': 1.474787} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.453839] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1141.454065] env[61936]: INFO nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Took 8.78 seconds to spawn the instance on the hypervisor. [ 1141.454257] env[61936]: DEBUG nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1141.455016] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313d14b2-6724-414e-b747-e46d46aec77d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.669240] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: d410e029-cb32-4cc6-b11d-9cba643e3b7b] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1141.683969] env[61936]: DEBUG oslo_vmware.api [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253592, 'name': ReconfigVM_Task, 'duration_secs': 5.721892} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.684217] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.684429] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Reconfigured VM to detach interface {{(pid=61936) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1141.948312] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.948614] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.949250] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.949463] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.949639] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.951731] env[61936]: INFO nova.compute.manager [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Terminating instance [ 1141.973549] env[61936]: INFO nova.compute.manager [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Took 13.65 seconds to build instance. [ 1142.172438] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 7ca9ee3d-1d1d-4ec3-9e69-8d1cc4725377] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1142.455218] env[61936]: DEBUG nova.compute.manager [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1142.455564] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1142.456433] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e67b13-5e31-4aee-9084-5c821b2b7cfd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.464349] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1142.464584] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c5f553e-ae1f-49bd-9e52-c1f3723aaa30 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.470060] env[61936]: DEBUG oslo_vmware.api [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1142.470060] env[61936]: value = "task-1253601" [ 1142.470060] env[61936]: _type = "Task" [ 1142.470060] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.478995] env[61936]: DEBUG oslo_concurrency.lockutils [None req-29a2ed01-a42f-4063-b703-05b831cc6726 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.158s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.479325] env[61936]: DEBUG oslo_vmware.api [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253601, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.523391] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.523868] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.523868] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.524031] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.524228] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.526316] env[61936]: INFO nova.compute.manager [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Terminating instance [ 1142.570679] env[61936]: DEBUG nova.compute.manager [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Received event network-changed-2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1142.570889] env[61936]: DEBUG nova.compute.manager [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Refreshing instance network info cache due to event network-changed-2d477dba-b68d-487f-89a8-216750d10485. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1142.571114] env[61936]: DEBUG oslo_concurrency.lockutils [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] Acquiring lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.571256] env[61936]: DEBUG oslo_concurrency.lockutils [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] Acquired lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.571414] env[61936]: DEBUG nova.network.neutron [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Refreshing network info cache for port 2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1142.675883] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 301b0504-5cf9-44e0-bd3e-342f8a89278b] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1142.975962] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.976162] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquired lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.976276] env[61936]: DEBUG nova.network.neutron [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1142.981071] env[61936]: DEBUG oslo_vmware.api [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253601, 'name': PowerOffVM_Task, 'duration_secs': 0.185429} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.981535] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1142.981663] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1142.981914] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75a41102-1285-4215-a3e6-0a782f979f8e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.030022] env[61936]: DEBUG nova.compute.manager [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1143.030251] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1143.031051] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6dbb152-4bc4-4197-a845-e49cb2562fd6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.038261] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.038549] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15a75d3a-0652-4f10-a18e-8541d42d7e5d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.044091] env[61936]: DEBUG oslo_vmware.api [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1143.044091] env[61936]: value = "task-1253603" [ 1143.044091] env[61936]: _type = "Task" [ 1143.044091] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.047911] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1143.048127] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1143.048307] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleting the datastore file [datastore1] b8e0c45d-341e-4ad4-8941-4532ebc53e40 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.048829] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44a81667-703a-4805-9f18-35a5a100bd70 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.053574] env[61936]: DEBUG oslo_vmware.api [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.057062] env[61936]: DEBUG oslo_vmware.api [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for the task: (returnval){ [ 1143.057062] env[61936]: value = "task-1253604" [ 1143.057062] env[61936]: _type = "Task" [ 1143.057062] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.065303] env[61936]: DEBUG oslo_vmware.api [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.179432] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 867a4bdc-8326-48b3-ba17-ca8336533c2d] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1143.505812] env[61936]: DEBUG nova.network.neutron [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updated VIF entry in instance network info cache for port 2d477dba-b68d-487f-89a8-216750d10485. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1143.506220] env[61936]: DEBUG nova.network.neutron [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updating instance_info_cache with network_info: [{"id": "2d477dba-b68d-487f-89a8-216750d10485", "address": "fa:16:3e:7d:5b:4b", "network": {"id": "fa124382-3bb7-4ec1-9d7b-35ed8ab0b392", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-43975413-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9ca963bd69594788806a9900942f895b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd72ef32-a57c-43b0-93df-e8a030987d44", "external-id": "nsx-vlan-transportzone-340", "segmentation_id": 340, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d477dba-b6", "ovs_interfaceid": "2d477dba-b68d-487f-89a8-216750d10485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.556728] env[61936]: DEBUG oslo_vmware.api [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253603, 'name': PowerOffVM_Task, 'duration_secs': 0.210723} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.556728] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1143.556908] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1143.557213] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39b1d56a-45aa-48ad-af07-8f3d0c69aa38 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.566486] env[61936]: DEBUG oslo_vmware.api [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Task: {'id': task-1253604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139989} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.566719] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.566907] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1143.567125] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1143.567291] env[61936]: INFO nova.compute.manager [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1143.567526] env[61936]: DEBUG oslo.service.loopingcall [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.567725] env[61936]: DEBUG nova.compute.manager [-] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1143.567809] env[61936]: DEBUG nova.network.neutron [-] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1143.619451] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1143.619808] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1143.620090] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleting the datastore file [datastore1] ea74d4a3-556f-4290-93ea-7f55f8c8bc2c {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.620455] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9468ef2b-8f77-4c60-844e-745dbc375c35 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.626819] env[61936]: DEBUG oslo_vmware.api [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1143.626819] env[61936]: value = "task-1253606" [ 1143.626819] env[61936]: _type = "Task" [ 1143.626819] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.634830] env[61936]: DEBUG oslo_vmware.api [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253606, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.683273] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 144ddda8-db07-4308-8fd3-4045067b5b37] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1143.713251] env[61936]: INFO nova.network.neutron [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Port 7e49da52-d495-432c-8a85-36a4301edcd1 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1143.713648] env[61936]: DEBUG nova.network.neutron [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [{"id": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "address": "fa:16:3e:e7:d1:79", "network": {"id": "03166185-d5e0-477b-892e-20bea284c8db", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1005180600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b9b0b8e03e6540a48947eebb51ad1135", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6ffea58a-e043-4ad1-9ad7-45a61fdca001", "external-id": "nsx-vlan-transportzone-662", "segmentation_id": 662, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef5dde7f-60", "ovs_interfaceid": "ef5dde7f-60d3-40b7-877c-53c75f17e14e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.009423] env[61936]: DEBUG oslo_concurrency.lockutils [req-a9deb84c-ce1e-476d-b605-cadfe4460c63 req-7a179a27-17b7-41eb-ab9c-a2a3da9bf297 service nova] Releasing lock "refresh_cache-2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.073811] env[61936]: DEBUG oslo_concurrency.lockutils [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.074128] env[61936]: DEBUG oslo_concurrency.lockutils [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.137849] env[61936]: DEBUG oslo_vmware.api [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253606, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146368} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.138115] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.138358] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1144.138615] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1144.138800] env[61936]: INFO nova.compute.manager [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1144.139074] env[61936]: DEBUG oslo.service.loopingcall [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1144.139290] env[61936]: DEBUG nova.compute.manager [-] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1144.139441] env[61936]: DEBUG nova.network.neutron [-] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1144.186089] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 1251e7f5-684c-428f-9d63-60b77084d09e] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1144.216466] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Releasing lock "refresh_cache-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.577575] env[61936]: DEBUG nova.compute.utils [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1144.599149] env[61936]: DEBUG nova.compute.manager [req-9e0b6765-3a4d-4a4a-9cdf-5efd3fd62b30 req-8f992261-11b9-43e6-8569-8da02db384bc service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Received event network-vif-deleted-92b2f7ae-8a1a-4908-939d-f2b01d922c5f {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1144.599149] env[61936]: INFO nova.compute.manager [req-9e0b6765-3a4d-4a4a-9cdf-5efd3fd62b30 req-8f992261-11b9-43e6-8569-8da02db384bc service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Neutron deleted interface 92b2f7ae-8a1a-4908-939d-f2b01d922c5f; detaching it from the instance and deleting it from the info cache [ 1144.599149] env[61936]: DEBUG nova.network.neutron [req-9e0b6765-3a4d-4a4a-9cdf-5efd3fd62b30 req-8f992261-11b9-43e6-8569-8da02db384bc service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.611832] env[61936]: DEBUG nova.network.neutron [-] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.688919] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 50cb3f79-39b2-46ab-a35c-fdad3bb266cd] Instance has had 0 of 5 cleanup attempts {{(pid=61936) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11639}} [ 1144.723072] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2b68b85e-4398-4cbf-ae21-9d34b96b6c69 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "interface-ea74d4a3-556f-4290-93ea-7f55f8c8bc2c-7e49da52-d495-432c-8a85-36a4301edcd1" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.624s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.081018] env[61936]: DEBUG oslo_concurrency.lockutils [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.101781] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68aac3f4-a118-4bb8-ae2e-595cc5cc80d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.111026] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee313f5-78e5-4631-9c3f-859a5b83db4a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.121600] env[61936]: INFO nova.compute.manager [-] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Took 1.55 seconds to deallocate network for instance. [ 1145.138650] env[61936]: DEBUG nova.compute.manager [req-9e0b6765-3a4d-4a4a-9cdf-5efd3fd62b30 req-8f992261-11b9-43e6-8569-8da02db384bc service nova] [instance: b8e0c45d-341e-4ad4-8941-4532ebc53e40] Detach interface failed, port_id=92b2f7ae-8a1a-4908-939d-f2b01d922c5f, reason: Instance b8e0c45d-341e-4ad4-8941-4532ebc53e40 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1145.192341] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.192691] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Cleaning up deleted instances with incomplete migration {{(pid=61936) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11664}} [ 1145.319574] env[61936]: DEBUG nova.network.neutron [-] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.628221] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.628641] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.628910] env[61936]: DEBUG nova.objects.instance [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lazy-loading 'resources' on Instance uuid b8e0c45d-341e-4ad4-8941-4532ebc53e40 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.695234] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1145.821502] env[61936]: INFO nova.compute.manager [-] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Took 1.68 seconds to deallocate network for instance. [ 1146.144425] env[61936]: DEBUG oslo_concurrency.lockutils [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.144593] env[61936]: DEBUG oslo_concurrency.lockutils [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.144827] env[61936]: INFO nova.compute.manager [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Attaching volume 30eb2cf1-2c87-40b1-8737-3d31a93c5cfc to /dev/sdb [ 1146.177645] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c41a89e-f6c7-405f-864e-919472b2b6f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.186901] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104e1942-cc1c-4c19-8222-5b79ccc997b8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.200515] env[61936]: DEBUG nova.virt.block_device [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updating existing volume attachment record: cf431755-0206-4f6a-8cf4-6e30c4f9cd1f {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1146.239024] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94df5dbf-b5c9-44b4-835c-37fee694a84e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.246507] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b41bf5-61ab-4594-94c3-a10202f720bc {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.276385] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b88480-ae71-4526-83d7-0c148ccb208e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.283372] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4a6b39-c273-4dd6-92c2-33b4a735ca90 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.296292] env[61936]: DEBUG nova.compute.provider_tree [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.328132] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.627482] env[61936]: DEBUG nova.compute.manager [req-795b9aa9-d848-4dd4-a508-4a10e5a014ee req-953ae771-92b1-4c70-b135-852c34983d0e service nova] [instance: ea74d4a3-556f-4290-93ea-7f55f8c8bc2c] Received event network-vif-deleted-ef5dde7f-60d3-40b7-877c-53c75f17e14e {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1146.799016] env[61936]: DEBUG nova.scheduler.client.report [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1147.198115] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.198352] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.304298] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.676s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.306586] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.979s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.306820] env[61936]: DEBUG nova.objects.instance [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'resources' on Instance uuid ea74d4a3-556f-4290-93ea-7f55f8c8bc2c {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.323740] env[61936]: INFO nova.scheduler.client.report [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Deleted allocations for instance b8e0c45d-341e-4ad4-8941-4532ebc53e40 [ 1147.704489] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.704489] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 1147.704489] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Rebuilding the list of instances to heal {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10330}} [ 1147.830564] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b1380c95-5292-4b95-9e24-06150d01acb4 tempest-AttachVolumeNegativeTest-557620526 tempest-AttachVolumeNegativeTest-557620526-project-member] Lock "b8e0c45d-341e-4ad4-8941-4532ebc53e40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.882s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.886425] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e2b334-0148-4dac-aa5d-206f7693141a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.894339] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7fd105-9f4c-493b-8bf1-31ace3e98f2b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.923533] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dde658a-6c32-45c3-8fe8-a79d5bf96b84 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.930361] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80af688b-afa4-4b4a-8c00-8578fc331f67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.944355] env[61936]: DEBUG nova.compute.provider_tree [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.235156] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.235319] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquired lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.235468] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Forcefully refreshing network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1148.235617] env[61936]: DEBUG nova.objects.instance [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lazy-loading 'info_cache' on Instance uuid e6930eaf-bb63-4553-8cd2-f2eb62fca510 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1148.447126] env[61936]: DEBUG nova.scheduler.client.report [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.951980] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.974738] env[61936]: INFO nova.scheduler.client.report [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted allocations for instance ea74d4a3-556f-4290-93ea-7f55f8c8bc2c [ 1149.483076] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1dcff2a1-953e-479d-ba95-a5cfa087d929 tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "ea74d4a3-556f-4290-93ea-7f55f8c8bc2c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.959s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.025625] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updating instance_info_cache with network_info: [{"id": "855dd2af-6f16-4185-88e2-0022244dc30d", "address": "fa:16:3e:b1:24:f2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855dd2af-6f", "ovs_interfaceid": "855dd2af-6f16-4185-88e2-0022244dc30d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.292050] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.292314] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.292549] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.292737] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.292911] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.295064] env[61936]: INFO nova.compute.manager [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Terminating instance [ 1150.528094] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Releasing lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.528398] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updated the network info_cache for instance {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10397}} [ 1150.528538] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.528665] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.528813] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.528950] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.529110] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.529236] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 1150.529375] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1150.748021] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1150.748021] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1150.748813] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65955a3-804a-4e94-9409-db088fbb0d6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.766149] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e7859d-f415-4c73-b6af-d6f474a2bd2a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.790286] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc/volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.790555] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68c112ba-7c48-4eb5-92a0-256c9d7d486a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.803385] env[61936]: DEBUG nova.compute.manager [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1150.803654] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1150.804735] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18de5be-f332-4f2b-ad0c-f5866f44ab95 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.812079] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1150.813177] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-555862a9-1c4c-4235-bdee-e57638dbf056 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.814603] env[61936]: DEBUG oslo_vmware.api [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1150.814603] env[61936]: value = "task-1253612" [ 1150.814603] env[61936]: _type = "Task" [ 1150.814603] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.819117] env[61936]: DEBUG oslo_vmware.api [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1150.819117] env[61936]: value = "task-1253613" [ 1150.819117] env[61936]: _type = "Task" [ 1150.819117] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.822349] env[61936]: DEBUG oslo_vmware.api [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.829352] env[61936]: DEBUG oslo_vmware.api [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253613, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.032205] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.032511] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.032709] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.032951] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1151.033843] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a668f7-632a-478f-abc9-f7a931110483 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.042163] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c950236-6436-4d06-bce8-ae0eb3fe2378 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.055879] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e2d04d-5db9-4e57-ac3f-d2a0df1a3e67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.062720] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2c25fd-1c62-45a7-ad1d-0ebacb0c471d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.092530] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180121MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1151.092672] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.092856] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.324152] env[61936]: DEBUG oslo_vmware.api [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253612, 'name': ReconfigVM_Task, 'duration_secs': 0.364623} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.326906] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc/volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1151.331541] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70be6e21-4017-4846-8806-9ca87b89ddaa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.345023] env[61936]: DEBUG oslo_vmware.api [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253613, 'name': PowerOffVM_Task, 'duration_secs': 0.176195} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.346110] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1151.346299] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1151.346585] env[61936]: DEBUG oslo_vmware.api [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1151.346585] env[61936]: value = "task-1253614" [ 1151.346585] env[61936]: _type = "Task" [ 1151.346585] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.346763] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-786bad57-72b9-4b5a-8f02-983a3d554462 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.355219] env[61936]: DEBUG oslo_vmware.api [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.411768] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1151.412062] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1151.412289] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleting the datastore file [datastore1] f1d45dcb-768d-447e-be49-ef43ad98fa61 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1151.412578] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea2e9885-247c-488c-bf98-45469f18508b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.418210] env[61936]: DEBUG oslo_vmware.api [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for the task: (returnval){ [ 1151.418210] env[61936]: value = "task-1253616" [ 1151.418210] env[61936]: _type = "Task" [ 1151.418210] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.425679] env[61936]: DEBUG oslo_vmware.api [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253616, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.857962] env[61936]: DEBUG oslo_vmware.api [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253614, 'name': ReconfigVM_Task, 'duration_secs': 0.154817} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.858293] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1151.928082] env[61936]: DEBUG oslo_vmware.api [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Task: {'id': task-1253616, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175311} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.929571] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1151.929571] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1151.929571] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1151.929571] env[61936]: INFO nova.compute.manager [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1151.929571] env[61936]: DEBUG oslo.service.loopingcall [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1151.929908] env[61936]: DEBUG nova.compute.manager [-] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1151.929908] env[61936]: DEBUG nova.network.neutron [-] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1152.120169] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e6930eaf-bb63-4553-8cd2-f2eb62fca510 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.120345] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance f1d45dcb-768d-447e-be49-ef43ad98fa61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.120464] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 3e337ad6-194b-4cd5-bfee-708ac00b7430 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.120582] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 767a9a43-784e-4270-92c7-53b55122dd8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.120696] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1152.120937] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1152.121040] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1152.169103] env[61936]: DEBUG nova.compute.manager [req-a831bfec-08ff-4418-9bf1-b9fbb17c41a3 req-57a452a3-f824-4963-8bc7-626e8c0d8b3b service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Received event network-vif-deleted-4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1152.169366] env[61936]: INFO nova.compute.manager [req-a831bfec-08ff-4418-9bf1-b9fbb17c41a3 req-57a452a3-f824-4963-8bc7-626e8c0d8b3b service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Neutron deleted interface 4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d; detaching it from the instance and deleting it from the info cache [ 1152.169577] env[61936]: DEBUG nova.network.neutron [req-a831bfec-08ff-4418-9bf1-b9fbb17c41a3 req-57a452a3-f824-4963-8bc7-626e8c0d8b3b service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.194310] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b16474c-978a-492b-a627-e4d5bf47480c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.201840] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-838a9ac5-68cf-4a6d-970e-de9e843e66a5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.232469] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac9857b-cbfb-4527-8346-a09f124f843b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.239629] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e83d55c-b392-4dc5-b85a-6961e2a62a10 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.252194] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.646792] env[61936]: DEBUG nova.network.neutron [-] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.672221] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e77c181-53f3-4dd2-8451-5999e2aadbd9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.681277] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e9b62c-82bc-44d4-be34-f3dbcfc5aa32 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.705732] env[61936]: DEBUG nova.compute.manager [req-a831bfec-08ff-4418-9bf1-b9fbb17c41a3 req-57a452a3-f824-4963-8bc7-626e8c0d8b3b service nova] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Detach interface failed, port_id=4e9fc5fb-74ae-4cf4-91ee-ee4ba897117d, reason: Instance f1d45dcb-768d-447e-be49-ef43ad98fa61 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1152.755198] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1152.894103] env[61936]: DEBUG nova.objects.instance [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'flavor' on Instance uuid 3e337ad6-194b-4cd5-bfee-708ac00b7430 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.149760] env[61936]: INFO nova.compute.manager [-] [instance: f1d45dcb-768d-447e-be49-ef43ad98fa61] Took 1.22 seconds to deallocate network for instance. [ 1153.258986] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1153.259198] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.166s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.398912] env[61936]: DEBUG oslo_concurrency.lockutils [None req-097e2f41-7df0-42ad-8ab0-aadbfac6d96e tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.254s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.656158] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.656458] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.656686] env[61936]: DEBUG nova.objects.instance [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lazy-loading 'resources' on Instance uuid f1d45dcb-768d-447e-be49-ef43ad98fa61 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.748398] env[61936]: INFO nova.compute.manager [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Rebuilding instance [ 1153.790543] env[61936]: DEBUG nova.compute.manager [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1153.791410] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e966f4b4-5ac8-4e08-82a6-4426cddffd67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.221905] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e01f05b-c331-4b34-b1a2-ff4c9766e3cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.229689] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68ede7a-e230-49ce-8f6e-ab58d33b9ca3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.259220] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bda594-c468-46d9-9b77-8576b765f75f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.266998] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b435cdd-70a1-44f4-9009-f851c05e2b09 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.280166] env[61936]: DEBUG nova.compute.provider_tree [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.782834] env[61936]: DEBUG nova.scheduler.client.report [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1154.803207] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1154.803538] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c86f06e4-6c11-48a5-b7c4-000a45717acb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.810637] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1154.810637] env[61936]: value = "task-1253618" [ 1154.810637] env[61936]: _type = "Task" [ 1154.810637] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.818949] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253618, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.288035] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.308923] env[61936]: INFO nova.scheduler.client.report [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Deleted allocations for instance f1d45dcb-768d-447e-be49-ef43ad98fa61 [ 1155.325183] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253618, 'name': PowerOffVM_Task, 'duration_secs': 0.230735} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.325570] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1155.374271] env[61936]: INFO nova.compute.manager [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Detaching volume 30eb2cf1-2c87-40b1-8737-3d31a93c5cfc [ 1155.403455] env[61936]: INFO nova.virt.block_device [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Attempting to driver detach volume 30eb2cf1-2c87-40b1-8737-3d31a93c5cfc from mountpoint /dev/sdb [ 1155.403701] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1155.403886] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1155.404770] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d554b11-7a6b-4e6b-ab89-ce3d84acf9f0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.427606] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a379c5-4712-4b88-94e5-2e0782d67450 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.435264] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21b5e34-86f5-43af-972d-63ba1774741e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.455227] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d098b1-92b4-4985-8d23-f93e41295813 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.470596] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] The volume has not been displaced from its original location: [datastore1] volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc/volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1155.476810] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1155.476810] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62063e04-5b62-4746-8784-8fe08ac47871 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.494385] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1155.494385] env[61936]: value = "task-1253620" [ 1155.494385] env[61936]: _type = "Task" [ 1155.494385] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.504363] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253620, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.821839] env[61936]: DEBUG oslo_concurrency.lockutils [None req-f9bbf0ef-1b79-438e-a7d8-7de6689eabfb tempest-AttachInterfacesTestJSON-1208629438 tempest-AttachInterfacesTestJSON-1208629438-project-member] Lock "f1d45dcb-768d-447e-be49-ef43ad98fa61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.529s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.004725] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253620, 'name': ReconfigVM_Task, 'duration_secs': 0.259136} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.004998] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1156.009582] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71c8a0d7-e3f4-4c19-9513-c87159d29337 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.024413] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1156.024413] env[61936]: value = "task-1253621" [ 1156.024413] env[61936]: _type = "Task" [ 1156.024413] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.032103] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253621, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.533648] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253621, 'name': ReconfigVM_Task, 'duration_secs': 0.480114} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.534089] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1157.581256] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.581702] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd482ca3-b94c-4401-a099-ed098dc8b151 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.589878] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1157.589878] env[61936]: value = "task-1253622" [ 1157.589878] env[61936]: _type = "Task" [ 1157.589878] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.598373] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253622, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.099683] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1158.099904] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1158.100127] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1158.100938] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f33d2fb-cf65-45be-b572-08b8e0296bed {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.120108] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12dc2549-425d-4fd4-8c2a-1f5247e38ca8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.125994] env[61936]: WARNING nova.virt.vmwareapi.driver [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1158.126277] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1158.126968] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e53ad0-9320-4f73-874e-af6a58b75989 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.133014] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1158.133210] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37b273ec-3df2-4f17-884c-60066c72f795 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.214299] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1158.214545] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1158.214730] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleting the datastore file [datastore2] 3e337ad6-194b-4cd5-bfee-708ac00b7430 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.215014] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b96c6af7-92a8-481d-8e89-cada0044fb18 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.222736] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1158.222736] env[61936]: value = "task-1253624" [ 1158.222736] env[61936]: _type = "Task" [ 1158.222736] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.231042] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253624, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.733477] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253624, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13504} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.733751] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1158.734246] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1158.734246] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1159.242728] env[61936]: INFO nova.virt.block_device [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Booting with volume 30eb2cf1-2c87-40b1-8737-3d31a93c5cfc at /dev/sdb [ 1159.284170] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c48aff2-06b9-425e-8fb4-c61bb4fdf33b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.294588] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0534acb5-c65c-454f-a4cb-afb962bbe4ef {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.321106] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ec4dc3d-3e42-4d73-8b61-46a4bdc661d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.329639] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6f13ee-4808-4805-8ae1-0ab87afdcdcb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.354436] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da94b962-4a0d-4dd5-973f-7599261131e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.361162] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15eff3f-39a9-4b71-8330-f9c4312e3a6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.374165] env[61936]: DEBUG nova.virt.block_device [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updating existing volume attachment record: 6f19f090-13cf-46cc-a6ab-8a73e57739ee {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1160.917715] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_power_states {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.422113] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Getting list of instances from cluster (obj){ [ 1161.422113] env[61936]: value = "domain-c8" [ 1161.422113] env[61936]: _type = "ClusterComputeResource" [ 1161.422113] env[61936]: } {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1161.423246] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb204677-84f0-402a-8c27-49c6bbf38f70 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.435423] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Got total of 3 instances {{(pid=61936) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1161.435587] env[61936]: WARNING nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] While synchronizing instance power states, found 4 instances in the database and 3 instances on the hypervisor. [ 1161.435724] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Triggering sync for uuid e6930eaf-bb63-4553-8cd2-f2eb62fca510 {{(pid=61936) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10736}} [ 1161.435915] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Triggering sync for uuid 3e337ad6-194b-4cd5-bfee-708ac00b7430 {{(pid=61936) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10736}} [ 1161.436089] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Triggering sync for uuid 767a9a43-784e-4270-92c7-53b55122dd8f {{(pid=61936) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10736}} [ 1161.436246] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Triggering sync for uuid 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 {{(pid=61936) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10736}} [ 1161.436569] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.436790] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.437059] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.437251] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.437421] env[61936]: INFO nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] During sync_power_state the instance has a pending task (rebuild_block_device_mapping). Skip. [ 1161.437586] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.437778] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.437956] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.438208] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.438391] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.439195] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7cdffb-f1e8-41cd-8f10-db5453031c55 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.442308] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e667551-0155-47fc-b3d1-bd4f38442afb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.444975] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491d96df-e297-40d8-abdc-3b27ded0c588 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.484443] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1161.484705] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1161.484865] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1161.485062] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1161.485215] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1161.485367] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1161.485574] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1161.485736] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1161.485904] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1161.486085] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1161.486266] env[61936]: DEBUG nova.virt.hardware [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1161.487405] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb279cca-0272-49ab-a018-338ce759fb31 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.495794] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e237da-f950-4e5c-a075-87183b87f992 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.510260] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:df:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ba0ab28-3350-4b44-9b20-cc64a712f113', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1161.517522] env[61936]: DEBUG oslo.service.loopingcall [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.517768] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1161.517982] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc7f193f-da74-4d91-a858-cd5cdca0d238 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.537318] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1161.537318] env[61936]: value = "task-1253625" [ 1161.537318] env[61936]: _type = "Task" [ 1161.537318] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.544681] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253625, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.957950] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.520s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.958373] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.959860] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.521s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.047628] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253625, 'name': CreateVM_Task, 'duration_secs': 0.313863} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.047789] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1162.048549] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.048644] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.048962] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1162.049534] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de556167-a1c9-4075-8a0f-b451d3fdfc14 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.054323] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1162.054323] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523f0b5b-8d0a-0e12-53fd-022f89796e6d" [ 1162.054323] env[61936]: _type = "Task" [ 1162.054323] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.062598] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523f0b5b-8d0a-0e12-53fd-022f89796e6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.565539] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]523f0b5b-8d0a-0e12-53fd-022f89796e6d, 'name': SearchDatastore_Task, 'duration_secs': 0.010315} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.565539] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.565796] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1162.566042] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.566196] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.566475] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1162.566808] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84e10589-194e-457f-80ad-3ae3ae55f99d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.574081] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.574296] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.574460] env[61936]: INFO nova.compute.manager [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Shelving [ 1162.578950] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1162.578950] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1162.578950] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99dd2939-2d64-4c90-94dc-0d11527e207e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.582752] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1162.582752] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5243b5c9-ea68-1f01-cbf5-2325fbb1cea1" [ 1162.582752] env[61936]: _type = "Task" [ 1162.582752] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.591521] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5243b5c9-ea68-1f01-cbf5-2325fbb1cea1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.094952] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]5243b5c9-ea68-1f01-cbf5-2325fbb1cea1, 'name': SearchDatastore_Task, 'duration_secs': 0.008849} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.095757] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7eab416-cb37-4795-aea5-b8020088bd41 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.100710] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1163.100710] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c898ba-7a5e-d4ff-c50a-7cf2486c4fef" [ 1163.100710] env[61936]: _type = "Task" [ 1163.100710] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.108379] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c898ba-7a5e-d4ff-c50a-7cf2486c4fef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.589763] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1163.590127] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd385fc6-eed3-428e-a666-6fbabb43804c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.597078] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1163.597078] env[61936]: value = "task-1253626" [ 1163.597078] env[61936]: _type = "Task" [ 1163.597078] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.273540] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253626, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.273885] env[61936]: WARNING oslo_vmware.common.loopingcall [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] task run outlasted interval by 0.17637000000000003 sec [ 1164.277367] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c898ba-7a5e-d4ff-c50a-7cf2486c4fef, 'name': SearchDatastore_Task, 'duration_secs': 0.008479} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.280349] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.280604] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1164.280842] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a242623e-ed01-4a6d-ba93-83da32319701 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.287711] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253626, 'name': PowerOffVM_Task, 'duration_secs': 0.174949} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.288746] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1164.289088] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1164.289088] env[61936]: value = "task-1253627" [ 1164.289088] env[61936]: _type = "Task" [ 1164.289088] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.289733] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b212cb2-cf43-4d02-b474-2e5a9b0d4e6d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.311874] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7224ff22-6d54-44e7-8d9f-48b07de01da8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.314198] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253627, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.800972] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253627, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440226} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.801265] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1164.801485] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1164.801755] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0488c3cd-cf00-4aff-903d-5a79ed720a08 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.807844] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1164.807844] env[61936]: value = "task-1253628" [ 1164.807844] env[61936]: _type = "Task" [ 1164.807844] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.815307] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253628, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.821634] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Creating Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1164.822031] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4cde7ceb-cb5b-4c27-bf52-53cf665c369e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.828185] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1164.828185] env[61936]: value = "task-1253629" [ 1164.828185] env[61936]: _type = "Task" [ 1164.828185] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.836814] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253629, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.318299] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253628, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071356} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.318662] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1165.319417] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257f2152-07de-4b77-875a-b50bb8f685a8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.341253] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1165.344115] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99529e5e-d14c-41cd-9a57-ef4958c6ca7a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.362674] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253629, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.363861] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1165.363861] env[61936]: value = "task-1253630" [ 1165.363861] env[61936]: _type = "Task" [ 1165.363861] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.370957] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253630, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.843446] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253629, 'name': CreateSnapshot_Task, 'duration_secs': 0.623355} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.844841] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Created Snapshot of the VM instance {{(pid=61936) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1165.844841] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293a03fa-0a29-43f1-ba9c-fd79988e0c3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.872110] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253630, 'name': ReconfigVM_Task, 'duration_secs': 0.295168} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.872450] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 3e337ad6-194b-4cd5-bfee-708ac00b7430/3e337ad6-194b-4cd5-bfee-708ac00b7430.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.873660] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'boot_index': 0, 'guest_format': None, 'encryption_options': None, 'encrypted': False, 'encryption_format': None, 'disk_bus': None, 'device_type': 'disk', 'image_id': '43acc3d3-5e18-42a0-9168-cb6831c6bbfb'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'}, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'disk_bus': None, 'device_type': None, 'attachment_id': '6f19f090-13cf-46cc-a6ab-8a73e57739ee', 'volume_type': None}], 'swap': None} {{(pid=61936) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1165.873978] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1165.874084] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1165.874830] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1cc6ce9-7c09-42d8-8f6c-085bdf03635e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.889629] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0843c92-4f10-491a-bd32-8ccf6b577a3e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.913019] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc/volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1165.913292] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0d3a6d9-9b28-447c-aca7-d2ca2265e662 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.929814] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1165.929814] env[61936]: value = "task-1253631" [ 1165.929814] env[61936]: _type = "Task" [ 1165.929814] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.937064] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253631, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.361546] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Creating linked-clone VM from snapshot {{(pid=61936) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1166.361918] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-31bac4d3-9469-4264-be76-802580c5e7ff {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.369511] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1166.369511] env[61936]: value = "task-1253632" [ 1166.369511] env[61936]: _type = "Task" [ 1166.369511] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.377204] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253632, 'name': CloneVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.440009] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253631, 'name': ReconfigVM_Task, 'duration_secs': 0.297644} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.440324] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc/volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1166.444999] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-177b9d35-f1f9-438d-98d8-1f48467b1388 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.459406] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1166.459406] env[61936]: value = "task-1253633" [ 1166.459406] env[61936]: _type = "Task" [ 1166.459406] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.467342] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.878925] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253632, 'name': CloneVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.968540] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253633, 'name': ReconfigVM_Task, 'duration_secs': 0.278724} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.968842] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1166.969481] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-47acfec6-e66a-430a-ac51-61713c0123f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.975551] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1166.975551] env[61936]: value = "task-1253634" [ 1166.975551] env[61936]: _type = "Task" [ 1166.975551] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.982999] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253634, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.379472] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253632, 'name': CloneVM_Task, 'duration_secs': 0.821169} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.379839] env[61936]: INFO nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Created linked-clone VM from snapshot [ 1167.380480] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e5d9e2-c53a-4559-a852-a20a5e54d5eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.388565] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Uploading image 72bb89e0-ac21-47b2-a15b-715a1930827c {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1167.413568] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1167.413568] env[61936]: value = "vm-270051" [ 1167.413568] env[61936]: _type = "VirtualMachine" [ 1167.413568] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1167.413845] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3475dc91-033b-4c0b-b08c-e8e6184e1e93 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.420271] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease: (returnval){ [ 1167.420271] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52053b34-99fc-bba9-3403-2e111cd9d2a9" [ 1167.420271] env[61936]: _type = "HttpNfcLease" [ 1167.420271] env[61936]: } obtained for exporting VM: (result){ [ 1167.420271] env[61936]: value = "vm-270051" [ 1167.420271] env[61936]: _type = "VirtualMachine" [ 1167.420271] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1167.420627] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the lease: (returnval){ [ 1167.420627] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52053b34-99fc-bba9-3403-2e111cd9d2a9" [ 1167.420627] env[61936]: _type = "HttpNfcLease" [ 1167.420627] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1167.426213] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1167.426213] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52053b34-99fc-bba9-3403-2e111cd9d2a9" [ 1167.426213] env[61936]: _type = "HttpNfcLease" [ 1167.426213] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1167.484546] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253634, 'name': Rename_Task, 'duration_secs': 0.152277} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.484857] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1167.485117] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3124b392-b54f-414a-bd07-9ca7e3a930f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.490264] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1167.490264] env[61936]: value = "task-1253636" [ 1167.490264] env[61936]: _type = "Task" [ 1167.490264] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.497091] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253636, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.929815] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1167.929815] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52053b34-99fc-bba9-3403-2e111cd9d2a9" [ 1167.929815] env[61936]: _type = "HttpNfcLease" [ 1167.929815] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1167.930164] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1167.930164] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52053b34-99fc-bba9-3403-2e111cd9d2a9" [ 1167.930164] env[61936]: _type = "HttpNfcLease" [ 1167.930164] env[61936]: }. {{(pid=61936) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1167.930887] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7467dc-3c5c-48b2-a7d9-f1aa0065b881 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.939284] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52631e55-537a-d311-4af2-c81dc01bf0ef/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1167.939574] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52631e55-537a-d311-4af2-c81dc01bf0ef/disk-0.vmdk for reading. {{(pid=61936) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1168.011779] env[61936]: DEBUG oslo_vmware.api [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253636, 'name': PowerOnVM_Task, 'duration_secs': 0.495633} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.012068] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1168.012338] env[61936]: DEBUG nova.compute.manager [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1168.013100] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b9a048-5a2e-4b65-b6cb-5f6fabb40b67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.038824] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e40d74fc-a200-4730-bb94-082643072e79 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.530755] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.531099] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.531339] env[61936]: DEBUG nova.objects.instance [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61936) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1169.539933] env[61936]: DEBUG oslo_concurrency.lockutils [None req-2cedf526-533c-40f1-baed-4a108b270b65 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.107608] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.107890] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.611237] env[61936]: INFO nova.compute.manager [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Detaching volume 30eb2cf1-2c87-40b1-8737-3d31a93c5cfc [ 1170.646208] env[61936]: INFO nova.virt.block_device [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Attempting to driver detach volume 30eb2cf1-2c87-40b1-8737-3d31a93c5cfc from mountpoint /dev/sdb [ 1170.646465] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1170.646652] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1170.647607] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80785c3-f4bb-4be5-a5b2-926fe4810402 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.669773] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6b8f3a-1818-4710-a31f-c7e0ad15916a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.676637] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c664d6f-15dc-42b3-86f8-517dea6d4d8a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.696474] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c07658-8bb7-4eb4-b732-8223e5a958f4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.710936] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] The volume has not been displaced from its original location: [datastore1] volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc/volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1170.716181] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1170.716522] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1c3b33f-d708-4df1-8e7b-764d13c52a8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.734446] env[61936]: DEBUG oslo_vmware.api [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1170.734446] env[61936]: value = "task-1253637" [ 1170.734446] env[61936]: _type = "Task" [ 1170.734446] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.742925] env[61936]: DEBUG oslo_vmware.api [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253637, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.244201] env[61936]: DEBUG oslo_vmware.api [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253637, 'name': ReconfigVM_Task, 'duration_secs': 0.255654} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.244484] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1171.248954] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12fbb424-1af2-4ab6-a1a6-96bc8cc555d8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.263517] env[61936]: DEBUG oslo_vmware.api [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1171.263517] env[61936]: value = "task-1253638" [ 1171.263517] env[61936]: _type = "Task" [ 1171.263517] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.271178] env[61936]: DEBUG oslo_vmware.api [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253638, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.773684] env[61936]: DEBUG oslo_vmware.api [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253638, 'name': ReconfigVM_Task, 'duration_secs': 0.271784} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.774093] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270048', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'name': 'volume-30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3e337ad6-194b-4cd5-bfee-708ac00b7430', 'attached_at': '', 'detached_at': '', 'volume_id': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc', 'serial': '30eb2cf1-2c87-40b1-8737-3d31a93c5cfc'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1172.316688] env[61936]: DEBUG nova.objects.instance [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'flavor' on Instance uuid 3e337ad6-194b-4cd5-bfee-708ac00b7430 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.324321] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1c8a9f34-b010-42f5-a5e8-ff73321b2dd0 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.216s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.359760] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.360154] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.360323] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "3e337ad6-194b-4cd5-bfee-708ac00b7430-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.360546] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.360731] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.362981] env[61936]: INFO nova.compute.manager [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Terminating instance [ 1174.867252] env[61936]: DEBUG nova.compute.manager [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1174.867520] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1174.868493] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e357f390-3647-4e32-a678-bc7dac91c5f3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.876401] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1174.876651] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9a15433-95ae-488d-b893-b0e266f7377b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.883088] env[61936]: DEBUG oslo_vmware.api [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1174.883088] env[61936]: value = "task-1253639" [ 1174.883088] env[61936]: _type = "Task" [ 1174.883088] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.891915] env[61936]: DEBUG oslo_vmware.api [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.393056] env[61936]: DEBUG oslo_vmware.api [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253639, 'name': PowerOffVM_Task, 'duration_secs': 0.193271} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.393493] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1175.393548] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1175.393764] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f711c1b3-53ce-4576-bedf-f1b313069898 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.462269] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1175.462468] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1175.462663] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleting the datastore file [datastore1] 3e337ad6-194b-4cd5-bfee-708ac00b7430 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1175.462938] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-756f6f44-4694-4356-8f96-041982ab959e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.470176] env[61936]: DEBUG oslo_vmware.api [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1175.470176] env[61936]: value = "task-1253641" [ 1175.470176] env[61936]: _type = "Task" [ 1175.470176] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.477819] env[61936]: DEBUG oslo_vmware.api [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.980150] env[61936]: DEBUG oslo_vmware.api [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253641, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.286625} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.980471] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1175.980607] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1175.980780] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1175.980949] env[61936]: INFO nova.compute.manager [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1175.981210] env[61936]: DEBUG oslo.service.loopingcall [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1175.981421] env[61936]: DEBUG nova.compute.manager [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1175.981490] env[61936]: DEBUG nova.network.neutron [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1176.473427] env[61936]: DEBUG nova.compute.manager [req-98ad384c-ef4f-4ab4-b791-8e36b5d15d0b req-fcc4f316-8bb8-48d4-b008-c3ab92f66076 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Received event network-vif-deleted-5ba0ab28-3350-4b44-9b20-cc64a712f113 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1176.474128] env[61936]: INFO nova.compute.manager [req-98ad384c-ef4f-4ab4-b791-8e36b5d15d0b req-fcc4f316-8bb8-48d4-b008-c3ab92f66076 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Neutron deleted interface 5ba0ab28-3350-4b44-9b20-cc64a712f113; detaching it from the instance and deleting it from the info cache [ 1176.474341] env[61936]: DEBUG nova.network.neutron [req-98ad384c-ef4f-4ab4-b791-8e36b5d15d0b req-fcc4f316-8bb8-48d4-b008-c3ab92f66076 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.901493] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52631e55-537a-d311-4af2-c81dc01bf0ef/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1176.902524] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6928939d-8797-4a7b-b407-66888e189ccd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.908939] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52631e55-537a-d311-4af2-c81dc01bf0ef/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1176.909124] env[61936]: ERROR oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52631e55-537a-d311-4af2-c81dc01bf0ef/disk-0.vmdk due to incomplete transfer. [ 1176.909348] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-dbee5aa5-ba3e-451d-ae9d-b7672f9cac38 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.916164] env[61936]: DEBUG oslo_vmware.rw_handles [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52631e55-537a-d311-4af2-c81dc01bf0ef/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1176.916364] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Uploaded image 72bb89e0-ac21-47b2-a15b-715a1930827c to the Glance image server {{(pid=61936) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1176.918556] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Destroying the VM {{(pid=61936) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1176.918800] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4845e27d-d5fe-4b0f-9438-4045e6ea29f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.924681] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1176.924681] env[61936]: value = "task-1253642" [ 1176.924681] env[61936]: _type = "Task" [ 1176.924681] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.933697] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253642, 'name': Destroy_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.952262] env[61936]: DEBUG nova.network.neutron [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.977126] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d720bc39-af92-4b31-b91a-a6786b12c543 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.986466] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da11abba-fc95-4f02-a3f6-0d122876eb99 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.012567] env[61936]: DEBUG nova.compute.manager [req-98ad384c-ef4f-4ab4-b791-8e36b5d15d0b req-fcc4f316-8bb8-48d4-b008-c3ab92f66076 service nova] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Detach interface failed, port_id=5ba0ab28-3350-4b44-9b20-cc64a712f113, reason: Instance 3e337ad6-194b-4cd5-bfee-708ac00b7430 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1177.434628] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253642, 'name': Destroy_Task, 'duration_secs': 0.299662} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.434963] env[61936]: INFO nova.virt.vmwareapi.vm_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Destroyed the VM [ 1177.435175] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Deleting Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1177.435424] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f71e630c-a346-4f94-a500-44d83d8bb564 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.441375] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1177.441375] env[61936]: value = "task-1253643" [ 1177.441375] env[61936]: _type = "Task" [ 1177.441375] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.448520] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253643, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.455262] env[61936]: INFO nova.compute.manager [-] [instance: 3e337ad6-194b-4cd5-bfee-708ac00b7430] Took 1.47 seconds to deallocate network for instance. [ 1177.951081] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253643, 'name': RemoveSnapshot_Task, 'duration_secs': 0.332863} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.951438] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Deleted Snapshot of the VM instance {{(pid=61936) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1177.951639] env[61936]: DEBUG nova.compute.manager [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1177.952389] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f751e75b-e967-4754-8f8c-22df05b9229f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.962826] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.963065] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.963279] env[61936]: DEBUG nova.objects.instance [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'resources' on Instance uuid 3e337ad6-194b-4cd5-bfee-708ac00b7430 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1178.463737] env[61936]: INFO nova.compute.manager [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Shelve offloading [ 1178.524712] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-479af337-d817-43ae-9f0d-6171eef0b378 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.532224] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1c6c8f-8d95-4286-8855-de11be3ad1c7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.562333] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce72725-221a-43ec-9fe4-9f7ffc4309d9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.569060] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9100f7bf-401a-4888-9740-fe3045d2dae2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.581469] env[61936]: DEBUG nova.compute.provider_tree [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.970025] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.970025] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-201419c8-11d0-4d2d-9789-031e58d300bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.976823] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1178.976823] env[61936]: value = "task-1253644" [ 1178.976823] env[61936]: _type = "Task" [ 1178.976823] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.984671] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253644, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.085031] env[61936]: DEBUG nova.scheduler.client.report [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1179.454876] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.455155] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.455305] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 1179.486947] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] VM already powered off {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1179.487170] env[61936]: DEBUG nova.compute.manager [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1179.487897] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba0e0e3-cec7-4195-9c60-4a66351e244b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.493152] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.493320] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.493497] env[61936]: DEBUG nova.network.neutron [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1179.589916] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.627s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.614475] env[61936]: INFO nova.scheduler.client.report [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted allocations for instance 3e337ad6-194b-4cd5-bfee-708ac00b7430 [ 1180.122211] env[61936]: DEBUG oslo_concurrency.lockutils [None req-a962b770-b80c-4791-add9-ffbc1be814bb tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "3e337ad6-194b-4cd5-bfee-708ac00b7430" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.762s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.411633] env[61936]: DEBUG nova.network.neutron [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77480d6a-c7", "ovs_interfaceid": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.915120] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.104144] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.104144] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.259099] env[61936]: DEBUG nova.compute.manager [req-56c5d291-2087-4df7-8cf6-68f8dd0db381 req-0372d7f3-e662-4a65-a524-2902f1306e38 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received event network-vif-unplugged-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1181.259378] env[61936]: DEBUG oslo_concurrency.lockutils [req-56c5d291-2087-4df7-8cf6-68f8dd0db381 req-0372d7f3-e662-4a65-a524-2902f1306e38 service nova] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.259531] env[61936]: DEBUG oslo_concurrency.lockutils [req-56c5d291-2087-4df7-8cf6-68f8dd0db381 req-0372d7f3-e662-4a65-a524-2902f1306e38 service nova] Lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.259698] env[61936]: DEBUG oslo_concurrency.lockutils [req-56c5d291-2087-4df7-8cf6-68f8dd0db381 req-0372d7f3-e662-4a65-a524-2902f1306e38 service nova] Lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.259861] env[61936]: DEBUG nova.compute.manager [req-56c5d291-2087-4df7-8cf6-68f8dd0db381 req-0372d7f3-e662-4a65-a524-2902f1306e38 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] No waiting events found dispatching network-vif-unplugged-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1181.262858] env[61936]: WARNING nova.compute.manager [req-56c5d291-2087-4df7-8cf6-68f8dd0db381 req-0372d7f3-e662-4a65-a524-2902f1306e38 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received unexpected event network-vif-unplugged-77480d6a-c7f5-41bf-9dc1-6554e54c4552 for instance with vm_state shelved and task_state shelving_offloading. [ 1181.317008] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1181.317947] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34708ad4-9f3d-4008-a9f8-c14db6843e84 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.325486] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1181.325722] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-573c3d26-a85a-4535-8c6b-bbcabcf9af2f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.393808] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1181.394052] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1181.394248] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleting the datastore file [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1181.394497] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae893574-cd5c-4702-a47c-fbadf7140223 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.400246] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1181.400246] env[61936]: value = "task-1253646" [ 1181.400246] env[61936]: _type = "Task" [ 1181.400246] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.407895] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253646, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.450670] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.455250] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.455405] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 1181.607061] env[61936]: DEBUG nova.compute.utils [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1181.909709] env[61936]: DEBUG oslo_vmware.api [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253646, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122232} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.909933] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1181.910134] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1181.910311] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1181.926184] env[61936]: INFO nova.scheduler.client.report [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted allocations for instance 767a9a43-784e-4270-92c7-53b55122dd8f [ 1182.109991] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.144039] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "149ac6e4-1233-4fec-8310-1794f9273b93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.144293] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "149ac6e4-1233-4fec-8310-1794f9273b93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.430292] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.430713] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.430860] env[61936]: DEBUG nova.objects.instance [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'resources' on Instance uuid 767a9a43-784e-4270-92c7-53b55122dd8f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1182.646949] env[61936]: DEBUG nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1182.933305] env[61936]: DEBUG nova.objects.instance [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'numa_topology' on Instance uuid 767a9a43-784e-4270-92c7-53b55122dd8f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.169314] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.190074] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.190074] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.190074] env[61936]: INFO nova.compute.manager [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Attaching volume 9f4a3886-dad1-440c-bbb2-a25fca51b49d to /dev/sdb [ 1183.218757] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6454d9c7-be5d-450f-a514-41041eb4a284 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.225549] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd62af2f-fe1b-4ce1-a1e8-68e8a12a108c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.238564] env[61936]: DEBUG nova.virt.block_device [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updating existing volume attachment record: da4fb442-8178-4752-9be1-8f216793bf2d {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1183.288700] env[61936]: DEBUG nova.compute.manager [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received event network-changed-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1183.288910] env[61936]: DEBUG nova.compute.manager [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Refreshing instance network info cache due to event network-changed-77480d6a-c7f5-41bf-9dc1-6554e54c4552. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1183.289186] env[61936]: DEBUG oslo_concurrency.lockutils [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] Acquiring lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.289336] env[61936]: DEBUG oslo_concurrency.lockutils [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] Acquired lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.289508] env[61936]: DEBUG nova.network.neutron [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Refreshing network info cache for port 77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1183.435994] env[61936]: DEBUG nova.objects.base [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Object Instance<767a9a43-784e-4270-92c7-53b55122dd8f> lazy-loaded attributes: resources,numa_topology {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1183.490027] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca76694-709a-46ed-8ad9-534cb3bcfc4f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.497359] env[61936]: INFO nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating ports in neutron [ 1183.500097] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4966fb29-8a34-4c92-bddd-4f71c121733c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.530448] env[61936]: INFO nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating port 77480d6a-c7f5-41bf-9dc1-6554e54c4552 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1183.532764] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505e7e66-3531-410e-8958-45dfaba85471 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.540172] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f8d6d4-f00e-41b3-b081-e022291789ac {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.554527] env[61936]: DEBUG nova.compute.provider_tree [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1184.046303] env[61936]: DEBUG nova.network.neutron [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updated VIF entry in instance network info cache for port 77480d6a-c7f5-41bf-9dc1-6554e54c4552. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1184.046762] env[61936]: DEBUG nova.network.neutron [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap77480d6a-c7", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.057553] env[61936]: DEBUG nova.scheduler.client.report [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1184.550025] env[61936]: DEBUG oslo_concurrency.lockutils [req-a673b3c9-7cd2-461c-b7fb-cfa160143a72 req-f4eeb4ad-5f3e-4f80-9cf2-0e362c3160fd service nova] Releasing lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.562084] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.131s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.564680] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.396s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.566062] env[61936]: INFO nova.compute.claims [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1184.895652] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.073985] env[61936]: DEBUG oslo_concurrency.lockutils [None req-1ec05c30-26c3-4a40-a609-9fd7ea5d3e47 tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.500s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.074821] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.179s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.075038] env[61936]: INFO nova.compute.manager [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Unshelving [ 1185.190501] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.190631] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquired lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.190817] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Forcefully refreshing network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1185.643188] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5ceae5-5b36-4585-bb59-3d6d6767f458 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.654767] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793368a6-9d7b-43ad-b598-e8d4e44d479b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.684732] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77a3785-306c-43ad-9424-79e0bdef6b26 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.691606] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fdbf2d1-3217-4dfa-bf05-36337eab43ca {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.706989] env[61936]: DEBUG nova.compute.provider_tree [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1186.095967] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.225883] env[61936]: ERROR nova.scheduler.client.report [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [req-2d8e9642-7b49-49ea-8878-44a87a76777c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID abd382a7-b427-4dcc-a050-3fb56fdd60ac. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2d8e9642-7b49-49ea-8878-44a87a76777c"}]} [ 1186.243077] env[61936]: DEBUG nova.scheduler.client.report [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Refreshing inventories for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1186.255656] env[61936]: DEBUG nova.scheduler.client.report [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating ProviderTree inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1186.255859] env[61936]: DEBUG nova.compute.provider_tree [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1186.266422] env[61936]: DEBUG nova.scheduler.client.report [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Refreshing aggregate associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, aggregates: None {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1186.283640] env[61936]: DEBUG nova.scheduler.client.report [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Refreshing trait associations for resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac, traits: COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61936) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1186.337160] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db32d221-0acc-47db-ae9b-8e5b2f328f5b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.347284] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47906c4-cce8-4d11-a5c5-b914302678b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.376676] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c71813-3d5e-46ec-8523-5a76e20e528c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.385405] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae1b449-4f27-452c-ab1e-f05ed1c7ac6e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.399033] env[61936]: DEBUG nova.compute.provider_tree [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1186.414367] env[61936]: DEBUG nova.network.neutron [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77480d6a-c7", "ovs_interfaceid": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.917368] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Releasing lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.917671] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updated the network info_cache for instance {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10397}} [ 1186.917736] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.917890] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.918056] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.918208] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.926956] env[61936]: DEBUG nova.scheduler.client.report [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 137 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1186.927237] env[61936]: DEBUG nova.compute.provider_tree [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 137 to 138 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1186.927420] env[61936]: DEBUG nova.compute.provider_tree [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1187.420574] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.432536] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.868s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.433112] env[61936]: DEBUG nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1187.435643] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.340s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.435852] env[61936]: DEBUG nova.objects.instance [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'pci_requests' on Instance uuid 767a9a43-784e-4270-92c7-53b55122dd8f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.778551] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1187.778803] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270052', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'name': 'volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'serial': '9f4a3886-dad1-440c-bbb2-a25fca51b49d'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1187.779729] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea6a0d6-fe4b-42e2-9653-e7f7a3f00913 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.796029] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeced476-005e-4775-9151-fe48433d4e00 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.819529] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d/volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1187.819765] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-433e53f1-46da-40d1-ac96-03c822e70bbb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.836978] env[61936]: DEBUG oslo_vmware.api [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1187.836978] env[61936]: value = "task-1253649" [ 1187.836978] env[61936]: _type = "Task" [ 1187.836978] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.844158] env[61936]: DEBUG oslo_vmware.api [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253649, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.938812] env[61936]: DEBUG nova.compute.utils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1187.941705] env[61936]: DEBUG nova.objects.instance [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lazy-loading 'numa_topology' on Instance uuid 767a9a43-784e-4270-92c7-53b55122dd8f {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.943046] env[61936]: DEBUG nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1187.943046] env[61936]: DEBUG nova.network.neutron [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1187.980436] env[61936]: DEBUG nova.policy [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f378ef8b8b5f4ae59c0d71dd1661bb59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa564b684410493fa0028fd345048e02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1188.235235] env[61936]: DEBUG nova.network.neutron [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Successfully created port: 545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1188.347293] env[61936]: DEBUG oslo_vmware.api [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253649, 'name': ReconfigVM_Task, 'duration_secs': 0.347878} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.347580] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d/volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.352460] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e6c3167-1527-4665-87da-683185897e83 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.367291] env[61936]: DEBUG oslo_vmware.api [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1188.367291] env[61936]: value = "task-1253650" [ 1188.367291] env[61936]: _type = "Task" [ 1188.367291] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.378111] env[61936]: DEBUG oslo_vmware.api [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253650, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.444023] env[61936]: DEBUG nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1188.447960] env[61936]: INFO nova.compute.claims [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1188.878532] env[61936]: DEBUG oslo_vmware.api [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253650, 'name': ReconfigVM_Task, 'duration_secs': 0.13312} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.878838] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270052', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'name': 'volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'serial': '9f4a3886-dad1-440c-bbb2-a25fca51b49d'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1189.454034] env[61936]: DEBUG nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T16:41:49Z,direct_url=,disk_format='vmdk',id=43acc3d3-5e18-42a0-9168-cb6831c6bbfb,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='0a2bb4cb48fe49aba25dc72f9f3195d6',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T16:41:50Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1189.482461] env[61936]: DEBUG nova.virt.hardware [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1189.483167] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a06dd68-b911-45b6-a3ab-ec801a976d60 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.493302] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102ccc7c-a803-4b4e-b23d-61a7beb3f15b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.519631] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5eb101-0c62-4271-be9a-40726de439f1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.526230] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f940c9cf-0e67-46b0-aa9e-66ecaa037fe1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.555117] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514ac21e-863f-4f85-ab35-07ed6ffbd390 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.562497] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e55c1e-9923-4440-8507-b8ae2b937483 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.576894] env[61936]: DEBUG nova.compute.provider_tree [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.695559] env[61936]: DEBUG nova.compute.manager [req-947614ef-3a0f-4f8a-81cc-03a316b4a469 req-8a669b27-85e1-474e-99c3-f06b0b52234f service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Received event network-vif-plugged-545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1189.695789] env[61936]: DEBUG oslo_concurrency.lockutils [req-947614ef-3a0f-4f8a-81cc-03a316b4a469 req-8a669b27-85e1-474e-99c3-f06b0b52234f service nova] Acquiring lock "149ac6e4-1233-4fec-8310-1794f9273b93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.695999] env[61936]: DEBUG oslo_concurrency.lockutils [req-947614ef-3a0f-4f8a-81cc-03a316b4a469 req-8a669b27-85e1-474e-99c3-f06b0b52234f service nova] Lock "149ac6e4-1233-4fec-8310-1794f9273b93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.696180] env[61936]: DEBUG oslo_concurrency.lockutils [req-947614ef-3a0f-4f8a-81cc-03a316b4a469 req-8a669b27-85e1-474e-99c3-f06b0b52234f service nova] Lock "149ac6e4-1233-4fec-8310-1794f9273b93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.696348] env[61936]: DEBUG nova.compute.manager [req-947614ef-3a0f-4f8a-81cc-03a316b4a469 req-8a669b27-85e1-474e-99c3-f06b0b52234f service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] No waiting events found dispatching network-vif-plugged-545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1189.696511] env[61936]: WARNING nova.compute.manager [req-947614ef-3a0f-4f8a-81cc-03a316b4a469 req-8a669b27-85e1-474e-99c3-f06b0b52234f service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Received unexpected event network-vif-plugged-545930bc-8fe4-41d2-8128-911d03a7bbfe for instance with vm_state building and task_state spawning. [ 1189.776291] env[61936]: DEBUG nova.network.neutron [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Successfully updated port: 545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1189.915622] env[61936]: DEBUG nova.objects.instance [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1190.080404] env[61936]: DEBUG nova.scheduler.client.report [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1190.279380] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.279544] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.279697] env[61936]: DEBUG nova.network.neutron [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1190.420393] env[61936]: DEBUG oslo_concurrency.lockutils [None req-b4fb9f52-75de-49e7-acd7-d2c993e5983e tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.231s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.585451] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.150s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.587505] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.167s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.587685] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.587841] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1190.588710] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef08e89f-cfd9-4201-9bb3-aa00d4b1102d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.596916] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc64ed6-6264-4dd1-8342-1c99b1a715cd {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.611854] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc68783-b691-4350-a3ab-c5f6160982a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.619488] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924fff04-f9cc-42cf-b34f-9897447f4ee5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.624402] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.624622] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.624835] env[61936]: DEBUG nova.network.neutron [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1190.654028] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180076MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1190.654028] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.654028] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.809193] env[61936]: DEBUG nova.network.neutron [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1190.926159] env[61936]: DEBUG nova.network.neutron [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Updating instance_info_cache with network_info: [{"id": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "address": "fa:16:3e:6d:1c:a2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap545930bc-8f", "ovs_interfaceid": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.254778] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.255107] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.379403] env[61936]: DEBUG nova.network.neutron [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [{"id": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "address": "fa:16:3e:c0:49:57", "network": {"id": "87e10aa9-63be-46e6-bba8-d4ddc96f6b03", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-617252384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "905821a1ff2b4011994c0d2d8bc08b13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77480d6a-c7", "ovs_interfaceid": "77480d6a-c7f5-41bf-9dc1-6554e54c4552", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.428739] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.429064] env[61936]: DEBUG nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Instance network_info: |[{"id": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "address": "fa:16:3e:6d:1c:a2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap545930bc-8f", "ovs_interfaceid": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1191.429496] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:1c:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '545930bc-8fe4-41d2-8128-911d03a7bbfe', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.437735] env[61936]: DEBUG oslo.service.loopingcall [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.437946] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.438185] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d694048-b762-4ee6-997d-c5d2a1c53f3b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.457867] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.457867] env[61936]: value = "task-1253651" [ 1191.457867] env[61936]: _type = "Task" [ 1191.457867] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.464971] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253651, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.681749] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e6930eaf-bb63-4553-8cd2-f2eb62fca510 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1191.682056] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1191.682103] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 149ac6e4-1233-4fec-8310-1794f9273b93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1191.682208] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 767a9a43-784e-4270-92c7-53b55122dd8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1191.682388] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1191.682557] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1191.727123] env[61936]: DEBUG nova.compute.manager [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Received event network-changed-545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1191.727246] env[61936]: DEBUG nova.compute.manager [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Refreshing instance network info cache due to event network-changed-545930bc-8fe4-41d2-8128-911d03a7bbfe. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1191.727429] env[61936]: DEBUG oslo_concurrency.lockutils [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] Acquiring lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.727630] env[61936]: DEBUG oslo_concurrency.lockutils [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] Acquired lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.727729] env[61936]: DEBUG nova.network.neutron [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Refreshing network info cache for port 545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1191.746551] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b16710-6c26-4a0d-93b3-617ff776340d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.755226] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb9fd92-ec19-4def-8d56-d524874383e2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.759801] env[61936]: DEBUG nova.compute.utils [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1191.792847] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff085a10-d67b-4b20-aff6-044b59c026fb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.800974] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc51cd1d-0d38-4ada-8a4d-7f52f03b84a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.815589] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.881867] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "refresh_cache-767a9a43-784e-4270-92c7-53b55122dd8f" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.908430] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3a0f0c4ccf04c6d4115439da308ac85e',container_format='bare',created_at=2024-10-10T16:53:43Z,direct_url=,disk_format='vmdk',id=72bb89e0-ac21-47b2-a15b-715a1930827c,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-63141669-shelved',owner='905821a1ff2b4011994c0d2d8bc08b13',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-10T16:53:58Z,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1191.908695] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.908855] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1191.909087] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.909246] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1191.909398] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1191.909638] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1191.909811] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1191.910199] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1191.910399] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1191.910694] env[61936]: DEBUG nova.virt.hardware [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1191.911551] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7af0a65-372e-4ded-95c9-895fcd8d43ab {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.919554] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f20a881-1234-4fa5-a824-fc27c4804ed5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.932651] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:49:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77480d6a-c7f5-41bf-9dc1-6554e54c4552', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.939804] env[61936]: DEBUG oslo.service.loopingcall [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.940052] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.940272] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-021af17b-daf6-4ae6-a444-e1f3f310374b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.958096] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.958096] env[61936]: value = "task-1253652" [ 1191.958096] env[61936]: _type = "Task" [ 1191.958096] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.967851] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253651, 'name': CreateVM_Task, 'duration_secs': 0.367345} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.970471] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1191.970651] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253652, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.971272] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.971439] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.971756] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1191.971988] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cebfe801-1cae-4e1c-93eb-516df4b813d6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.975941] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1191.975941] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f28a92-aa34-6391-ada1-753444b1dd45" [ 1191.975941] env[61936]: _type = "Task" [ 1191.975941] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.983636] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f28a92-aa34-6391-ada1-753444b1dd45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.261828] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.318374] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1192.413070] env[61936]: DEBUG nova.network.neutron [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Updated VIF entry in instance network info cache for port 545930bc-8fe4-41d2-8128-911d03a7bbfe. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1192.413441] env[61936]: DEBUG nova.network.neutron [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Updating instance_info_cache with network_info: [{"id": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "address": "fa:16:3e:6d:1c:a2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap545930bc-8f", "ovs_interfaceid": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.468923] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253652, 'name': CreateVM_Task, 'duration_secs': 0.302321} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.470052] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1192.470052] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.470207] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.470465] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1192.470736] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77860d19-a8b8-4c33-bfd7-3df9f64d72ad {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.474807] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1192.474807] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a9475e-40bc-e355-6d48-4f64160455a2" [ 1192.474807] env[61936]: _type = "Task" [ 1192.474807] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.484297] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52a9475e-40bc-e355-6d48-4f64160455a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.487506] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52f28a92-aa34-6391-ada1-753444b1dd45, 'name': SearchDatastore_Task, 'duration_secs': 0.012021} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.487781] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.488036] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Processing image 43acc3d3-5e18-42a0-9168-cb6831c6bbfb {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.488267] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.488410] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.488585] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.489124] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47b1ced6-2d55-4b86-acb1-6d627ebbaa90 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.504824] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.504999] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1192.505676] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e56fd163-4cb9-4e99-8983-d9a8a30c2e65 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.510569] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1192.510569] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c09159-c3f1-48c8-6711-52d0d9cf5049" [ 1192.510569] env[61936]: _type = "Task" [ 1192.510569] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.517915] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c09159-c3f1-48c8-6711-52d0d9cf5049, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.822887] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1192.823779] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.169s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.916234] env[61936]: DEBUG oslo_concurrency.lockutils [req-10593526-b415-485c-970d-27471103c1b8 req-f501dd9e-3936-43d8-99b3-def6b8027741 service nova] Releasing lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.985176] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.985438] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Processing image 72bb89e0-ac21-47b2-a15b-715a1930827c {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.985710] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c/72bb89e0-ac21-47b2-a15b-715a1930827c.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.985864] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquired lock "[datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c/72bb89e0-ac21-47b2-a15b-715a1930827c.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.986061] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.986312] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97a3d13b-62a5-4d2a-8917-9a996be1586a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.994056] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.994237] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61936) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1192.994933] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51365926-e796-4e9a-80ed-57cf63a337fa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.999425] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1192.999425] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529e8de5-7a14-edad-c9b5-6bd589273b91" [ 1192.999425] env[61936]: _type = "Task" [ 1192.999425] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.006479] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]529e8de5-7a14-edad-c9b5-6bd589273b91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.019187] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52c09159-c3f1-48c8-6711-52d0d9cf5049, 'name': SearchDatastore_Task, 'duration_secs': 0.007694} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.019823] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e371544-eaca-4e6c-8a1b-e04c59a2285d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.024289] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1193.024289] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bc36c6-b336-7c42-2e05-7dfa6b085608" [ 1193.024289] env[61936]: _type = "Task" [ 1193.024289] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.031030] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bc36c6-b336-7c42-2e05-7dfa6b085608, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.318992] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.319314] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.319560] env[61936]: INFO nova.compute.manager [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Attaching volume 09790134-e491-49c4-81b0-7fc70d8ea557 to /dev/sdc [ 1193.348959] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053151ca-4342-4c8a-bb68-3f9ed8ebd4a7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.356303] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9dc3a3-d62a-40aa-8b70-a61e8423b928 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.360072] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.368721] env[61936]: DEBUG nova.virt.block_device [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updating existing volume attachment record: 1b7b50de-8b8d-482a-be69-696661ceb216 {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1193.509489] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Preparing fetch location {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1193.509741] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Fetch image to [datastore1] OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8/OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8.vmdk {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1193.509926] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Downloading stream optimized image 72bb89e0-ac21-47b2-a15b-715a1930827c to [datastore1] OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8/OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8.vmdk on the data store datastore1 as vApp {{(pid=61936) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1193.510109] env[61936]: DEBUG nova.virt.vmwareapi.images [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Downloading image file data 72bb89e0-ac21-47b2-a15b-715a1930827c to the ESX as VM named 'OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8' {{(pid=61936) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1193.548538] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52bc36c6-b336-7c42-2e05-7dfa6b085608, 'name': SearchDatastore_Task, 'duration_secs': 0.015877} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.548801] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "[datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.549206] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 149ac6e4-1233-4fec-8310-1794f9273b93/149ac6e4-1233-4fec-8310-1794f9273b93.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1193.549558] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36b83cd1-7ce2-466a-b201-da820d90abd4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.555309] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1193.555309] env[61936]: value = "task-1253654" [ 1193.555309] env[61936]: _type = "Task" [ 1193.555309] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.567174] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253654, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.591077] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1193.591077] env[61936]: value = "resgroup-9" [ 1193.591077] env[61936]: _type = "ResourcePool" [ 1193.591077] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1193.591404] env[61936]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ce80a3c6-c5b0-4f98-934e-2e1d39e968a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.611434] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease: (returnval){ [ 1193.611434] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd8683-4ec8-a430-f7fa-3c94b2128d7e" [ 1193.611434] env[61936]: _type = "HttpNfcLease" [ 1193.611434] env[61936]: } obtained for vApp import into resource pool (val){ [ 1193.611434] env[61936]: value = "resgroup-9" [ 1193.611434] env[61936]: _type = "ResourcePool" [ 1193.611434] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1193.611839] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the lease: (returnval){ [ 1193.611839] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd8683-4ec8-a430-f7fa-3c94b2128d7e" [ 1193.611839] env[61936]: _type = "HttpNfcLease" [ 1193.611839] env[61936]: } to be ready. {{(pid=61936) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1193.618138] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1193.618138] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd8683-4ec8-a430-f7fa-3c94b2128d7e" [ 1193.618138] env[61936]: _type = "HttpNfcLease" [ 1193.618138] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1194.066841] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253654, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474607} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.067217] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/43acc3d3-5e18-42a0-9168-cb6831c6bbfb/43acc3d3-5e18-42a0-9168-cb6831c6bbfb.vmdk to [datastore1] 149ac6e4-1233-4fec-8310-1794f9273b93/149ac6e4-1233-4fec-8310-1794f9273b93.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1194.067413] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Extending root virtual disk to 1048576 {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1194.067779] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e46bea3-3a0b-46cc-8363-c9c149daec43 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.074925] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1194.074925] env[61936]: value = "task-1253656" [ 1194.074925] env[61936]: _type = "Task" [ 1194.074925] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.085404] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.119612] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1194.119612] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd8683-4ec8-a430-f7fa-3c94b2128d7e" [ 1194.119612] env[61936]: _type = "HttpNfcLease" [ 1194.119612] env[61936]: } is initializing. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1194.584053] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062435} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.584329] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Extended root virtual disk {{(pid=61936) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1194.585135] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b9d1c6-45ba-435a-b54e-4ff9a7e30c0e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.605931] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 149ac6e4-1233-4fec-8310-1794f9273b93/149ac6e4-1233-4fec-8310-1794f9273b93.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1194.606165] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6278c71d-4f5d-4966-8689-446308312aeb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.625908] env[61936]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1194.625908] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd8683-4ec8-a430-f7fa-3c94b2128d7e" [ 1194.625908] env[61936]: _type = "HttpNfcLease" [ 1194.625908] env[61936]: } is ready. {{(pid=61936) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1194.626155] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1194.626155] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]52dd8683-4ec8-a430-f7fa-3c94b2128d7e" [ 1194.626155] env[61936]: _type = "HttpNfcLease" [ 1194.626155] env[61936]: }. {{(pid=61936) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1194.627604] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cef876-0f36-494d-875d-a73fffe72d8c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.629768] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1194.629768] env[61936]: value = "task-1253657" [ 1194.629768] env[61936]: _type = "Task" [ 1194.629768] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.635503] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f14b86-0169-2069-0d52-e88dab1113ac/disk-0.vmdk from lease info. {{(pid=61936) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1194.635681] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f14b86-0169-2069-0d52-e88dab1113ac/disk-0.vmdk. {{(pid=61936) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1194.694414] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253657, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.702053] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8335a0f3-431c-4018-9afa-96caee1a2539 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.141621] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253657, 'name': ReconfigVM_Task, 'duration_secs': 0.267829} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.143447] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 149ac6e4-1233-4fec-8310-1794f9273b93/149ac6e4-1233-4fec-8310-1794f9273b93.vmdk or device None with type sparse {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1195.144198] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f759fe74-9e6f-4ffa-b870-cfd90f723e13 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.151303] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1195.151303] env[61936]: value = "task-1253658" [ 1195.151303] env[61936]: _type = "Task" [ 1195.151303] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.163384] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253658, 'name': Rename_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.663141] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253658, 'name': Rename_Task, 'duration_secs': 0.184228} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.665844] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1195.666195] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19c844d6-c85d-4150-98e3-a4f75dc9e9a2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.673326] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1195.673326] env[61936]: value = "task-1253660" [ 1195.673326] env[61936]: _type = "Task" [ 1195.673326] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.682687] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.789699] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Completed reading data from the image iterator. {{(pid=61936) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1195.790058] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f14b86-0169-2069-0d52-e88dab1113ac/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1195.791211] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1443a6-91f9-449d-b95b-a96fc686354a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.799488] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f14b86-0169-2069-0d52-e88dab1113ac/disk-0.vmdk is in state: ready. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1195.799705] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f14b86-0169-2069-0d52-e88dab1113ac/disk-0.vmdk. {{(pid=61936) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1195.799980] env[61936]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-fcdba730-cc91-4622-a7b5-1618ca760e58 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.990091] env[61936]: DEBUG oslo_vmware.rw_handles [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f14b86-0169-2069-0d52-e88dab1113ac/disk-0.vmdk. {{(pid=61936) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1195.990347] env[61936]: INFO nova.virt.vmwareapi.images [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Downloaded image file data 72bb89e0-ac21-47b2-a15b-715a1930827c [ 1195.991248] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a0fb54-e15d-4f0c-b070-08f37435b98b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.008506] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d0664b2-f4f9-4e5a-9f87-9dd22f26439d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.027786] env[61936]: INFO nova.virt.vmwareapi.images [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] The imported VM was unregistered [ 1196.030077] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Caching image {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1196.030302] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Creating directory with path [datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1196.030564] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c779a83f-365d-4898-a58b-f3a73c97d64b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.039877] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Created directory with path [datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c {{(pid=61936) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1196.040112] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8/OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8.vmdk to [datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c/72bb89e0-ac21-47b2-a15b-715a1930827c.vmdk. {{(pid=61936) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1196.040387] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0f29b0c1-fdfd-41c8-82fe-aafb0e404847 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.045987] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1196.045987] env[61936]: value = "task-1253662" [ 1196.045987] env[61936]: _type = "Task" [ 1196.045987] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.054336] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253662, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.185841] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253660, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.555887] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253662, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.685325] env[61936]: DEBUG oslo_vmware.api [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253660, 'name': PowerOnVM_Task, 'duration_secs': 0.56013} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.685621] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1196.685842] env[61936]: INFO nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Took 7.23 seconds to spawn the instance on the hypervisor. [ 1196.686025] env[61936]: DEBUG nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1196.686857] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a666a3ce-295a-4978-a1f6-589e02c216f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.058240] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253662, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.209612] env[61936]: INFO nova.compute.manager [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Took 14.06 seconds to build instance. [ 1197.376584] env[61936]: DEBUG nova.compute.manager [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Received event network-changed-545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1197.377833] env[61936]: DEBUG nova.compute.manager [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Refreshing instance network info cache due to event network-changed-545930bc-8fe4-41d2-8128-911d03a7bbfe. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1197.377833] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] Acquiring lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.377833] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] Acquired lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.377833] env[61936]: DEBUG nova.network.neutron [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Refreshing network info cache for port 545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1197.557512] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253662, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.710770] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ab32d3dd-b072-43f1-9766-34102747e37d tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "149ac6e4-1233-4fec-8310-1794f9273b93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.566s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.912026] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1197.912026] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270055', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'name': 'volume-09790134-e491-49c4-81b0-7fc70d8ea557', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'serial': '09790134-e491-49c4-81b0-7fc70d8ea557'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1197.912936] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab1e835-4b1c-4259-86d8-3737dbf7fed9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.932442] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29bdea3-6ae7-4287-91a5-278dd2a23991 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.962150] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-09790134-e491-49c4-81b0-7fc70d8ea557/volume-09790134-e491-49c4-81b0-7fc70d8ea557.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1197.962528] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a09003cf-1695-4fbe-b8d7-36f6c8ae86d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.981164] env[61936]: DEBUG oslo_vmware.api [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1197.981164] env[61936]: value = "task-1253663" [ 1197.981164] env[61936]: _type = "Task" [ 1197.981164] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.990556] env[61936]: DEBUG oslo_vmware.api [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253663, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.058395] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253662, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.150192] env[61936]: DEBUG nova.network.neutron [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Updated VIF entry in instance network info cache for port 545930bc-8fe4-41d2-8128-911d03a7bbfe. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1198.150608] env[61936]: DEBUG nova.network.neutron [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Updating instance_info_cache with network_info: [{"id": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "address": "fa:16:3e:6d:1c:a2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap545930bc-8f", "ovs_interfaceid": "545930bc-8fe4-41d2-8128-911d03a7bbfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.490813] env[61936]: DEBUG oslo_vmware.api [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253663, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.557868] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253662, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.49407} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.558159] env[61936]: INFO nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8/OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8.vmdk to [datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c/72bb89e0-ac21-47b2-a15b-715a1930827c.vmdk. [ 1198.558356] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Cleaning up location [datastore1] OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8 {{(pid=61936) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1198.558524] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_0482c4e5-704f-47da-8b9f-dde31593e4e8 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1198.558789] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f5016a3-826a-4387-807d-5ffdc2b87e39 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.564250] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1198.564250] env[61936]: value = "task-1253664" [ 1198.564250] env[61936]: _type = "Task" [ 1198.564250] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.571276] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253664, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.653485] env[61936]: DEBUG oslo_concurrency.lockutils [req-7a91d8d6-f5b4-4112-acfb-19e79031296a req-e95f7979-d109-49d0-9803-0ec01a52be9e service nova] Releasing lock "refresh_cache-149ac6e4-1233-4fec-8310-1794f9273b93" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.991771] env[61936]: DEBUG oslo_vmware.api [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253663, 'name': ReconfigVM_Task, 'duration_secs': 0.872148} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.992093] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-09790134-e491-49c4-81b0-7fc70d8ea557/volume-09790134-e491-49c4-81b0-7fc70d8ea557.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1198.998076] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8123ce27-bc94-44c9-9f30-99220139debe {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.015691] env[61936]: DEBUG oslo_vmware.api [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1199.015691] env[61936]: value = "task-1253666" [ 1199.015691] env[61936]: _type = "Task" [ 1199.015691] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.027584] env[61936]: DEBUG oslo_vmware.api [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253666, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.073133] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253664, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.049287} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.073424] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1199.073606] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Releasing lock "[datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c/72bb89e0-ac21-47b2-a15b-715a1930827c.vmdk" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.073910] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c/72bb89e0-ac21-47b2-a15b-715a1930827c.vmdk to [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1199.074210] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97937b50-bc6b-49d5-a45a-68db901ce247 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.080177] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1199.080177] env[61936]: value = "task-1253667" [ 1199.080177] env[61936]: _type = "Task" [ 1199.080177] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.087270] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253667, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.528777] env[61936]: DEBUG oslo_vmware.api [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253666, 'name': ReconfigVM_Task, 'duration_secs': 0.191755} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.529208] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270055', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'name': 'volume-09790134-e491-49c4-81b0-7fc70d8ea557', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'serial': '09790134-e491-49c4-81b0-7fc70d8ea557'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1199.589695] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253667, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.091321] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253667, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.571103] env[61936]: DEBUG nova.objects.instance [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1200.592594] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253667, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.876112] env[61936]: DEBUG oslo_concurrency.lockutils [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.076480] env[61936]: DEBUG oslo_concurrency.lockutils [None req-d97a828d-acef-4c47-96e9-d8f4df9a0a3c tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.757s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.077490] env[61936]: DEBUG oslo_concurrency.lockutils [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.202s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.092022] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253667, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.579915] env[61936]: INFO nova.compute.manager [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Detaching volume 9f4a3886-dad1-440c-bbb2-a25fca51b49d [ 1201.593367] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253667, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.614905] env[61936]: INFO nova.virt.block_device [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Attempting to driver detach volume 9f4a3886-dad1-440c-bbb2-a25fca51b49d from mountpoint /dev/sdb [ 1201.615241] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1201.615445] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270052', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'name': 'volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'serial': '9f4a3886-dad1-440c-bbb2-a25fca51b49d'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1201.616404] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5333422-7bdc-4a3a-9b59-6ebd35c7eb47 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.640996] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eff8ece-7e7d-4c29-9fe5-54a566dfff94 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.647998] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2343a4a8-87ae-462f-a6a6-5d8e4bf0eb67 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.670709] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67177a3-a28b-4460-b260-1e100d60734a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.686731] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] The volume has not been displaced from its original location: [datastore1] volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d/volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1201.691864] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1201.692170] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6161ee4f-4448-4c54-9ca1-27e370e6a824 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.709342] env[61936]: DEBUG oslo_vmware.api [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1201.709342] env[61936]: value = "task-1253669" [ 1201.709342] env[61936]: _type = "Task" [ 1201.709342] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.716781] env[61936]: DEBUG oslo_vmware.api [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253669, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.092187] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253667, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.662242} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.092445] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/72bb89e0-ac21-47b2-a15b-715a1930827c/72bb89e0-ac21-47b2-a15b-715a1930827c.vmdk to [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk {{(pid=61936) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1202.093203] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5476e0-fdb3-43b9-85ec-f0a20bba314b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.114519] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1202.114759] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee010bbc-4f7b-4dcb-adbe-5c84720245f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.133389] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1202.133389] env[61936]: value = "task-1253670" [ 1202.133389] env[61936]: _type = "Task" [ 1202.133389] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.140716] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253670, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.218989] env[61936]: DEBUG oslo_vmware.api [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253669, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.643793] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253670, 'name': ReconfigVM_Task, 'duration_secs': 0.330031} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.644215] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f/767a9a43-784e-4270-92c7-53b55122dd8f.vmdk or device None with type streamOptimized {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.644747] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2759a4e-cfbd-4320-b42c-1e194359f66b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.651307] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1202.651307] env[61936]: value = "task-1253671" [ 1202.651307] env[61936]: _type = "Task" [ 1202.651307] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.660089] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253671, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.724689] env[61936]: DEBUG oslo_vmware.api [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253669, 'name': ReconfigVM_Task, 'duration_secs': 0.804221} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.724994] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1202.729625] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5356c56-e40a-45af-9d09-e45f6eff182f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.746441] env[61936]: DEBUG oslo_vmware.api [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1202.746441] env[61936]: value = "task-1253672" [ 1202.746441] env[61936]: _type = "Task" [ 1202.746441] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.754847] env[61936]: DEBUG oslo_vmware.api [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253672, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.161645] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253671, 'name': Rename_Task, 'duration_secs': 0.157441} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.161924] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1203.162191] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2069c3d-10c7-4ffb-9d01-048647e144d5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.168250] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1203.168250] env[61936]: value = "task-1253673" [ 1203.168250] env[61936]: _type = "Task" [ 1203.168250] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.175608] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.255533] env[61936]: DEBUG oslo_vmware.api [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253672, 'name': ReconfigVM_Task, 'duration_secs': 0.291252} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.255865] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270052', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'name': 'volume-9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '9f4a3886-dad1-440c-bbb2-a25fca51b49d', 'serial': '9f4a3886-dad1-440c-bbb2-a25fca51b49d'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1203.678391] env[61936]: DEBUG oslo_vmware.api [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253673, 'name': PowerOnVM_Task, 'duration_secs': 0.495972} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.678804] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1203.774728] env[61936]: DEBUG nova.compute.manager [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1203.775859] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3b0c9b-2aa4-44fe-930e-6a8d2aa5083e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.797145] env[61936]: DEBUG nova.objects.instance [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.293717] env[61936]: DEBUG oslo_concurrency.lockutils [None req-01169927-de97-448b-b566-9b7fa53ef61b tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.219s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.805065] env[61936]: DEBUG oslo_concurrency.lockutils [None req-af9b5816-7fd3-45a7-acde-af9cb8db0152 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.727s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.162279] env[61936]: DEBUG oslo_concurrency.lockutils [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.162512] env[61936]: DEBUG oslo_concurrency.lockutils [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.665483] env[61936]: INFO nova.compute.manager [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Detaching volume 09790134-e491-49c4-81b0-7fc70d8ea557 [ 1205.697684] env[61936]: INFO nova.virt.block_device [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Attempting to driver detach volume 09790134-e491-49c4-81b0-7fc70d8ea557 from mountpoint /dev/sdc [ 1205.697945] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1205.698161] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270055', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'name': 'volume-09790134-e491-49c4-81b0-7fc70d8ea557', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'serial': '09790134-e491-49c4-81b0-7fc70d8ea557'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1205.699063] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a25f71-bad4-44d5-8077-892f99973c42 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.719871] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bb34a7-da8e-4044-a52e-67027a45dec7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.726408] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b26d75-69a5-4214-aa30-7d4aec6f47d4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.746065] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b7d99-5770-4bde-9e2d-78d768b6e5c2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.760661] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] The volume has not been displaced from its original location: [datastore1] volume-09790134-e491-49c4-81b0-7fc70d8ea557/volume-09790134-e491-49c4-81b0-7fc70d8ea557.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1205.765851] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfiguring VM instance instance-0000006e to detach disk 2002 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1205.766150] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bae7a74f-57fa-43c7-83f7-e82b52dc5609 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.782615] env[61936]: DEBUG oslo_vmware.api [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1205.782615] env[61936]: value = "task-1253676" [ 1205.782615] env[61936]: _type = "Task" [ 1205.782615] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.791493] env[61936]: DEBUG oslo_vmware.api [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253676, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.293592] env[61936]: DEBUG oslo_vmware.api [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253676, 'name': ReconfigVM_Task, 'duration_secs': 0.243743} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.293949] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Reconfigured VM instance instance-0000006e to detach disk 2002 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1206.298630] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6b5fded-cee4-45a9-86b6-401d2999f144 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.313429] env[61936]: DEBUG oslo_vmware.api [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1206.313429] env[61936]: value = "task-1253677" [ 1206.313429] env[61936]: _type = "Task" [ 1206.313429] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.321567] env[61936]: DEBUG oslo_vmware.api [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253677, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.822767] env[61936]: DEBUG oslo_vmware.api [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253677, 'name': ReconfigVM_Task, 'duration_secs': 0.129981} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.823179] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270055', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'name': 'volume-09790134-e491-49c4-81b0-7fc70d8ea557', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2c20a6cb-8d9b-4d4a-9c5e-3a3774292496', 'attached_at': '', 'detached_at': '', 'volume_id': '09790134-e491-49c4-81b0-7fc70d8ea557', 'serial': '09790134-e491-49c4-81b0-7fc70d8ea557'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1207.361419] env[61936]: DEBUG nova.objects.instance [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'flavor' on Instance uuid 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.368681] env[61936]: DEBUG oslo_concurrency.lockutils [None req-780faeba-a607-4e38-8fb1-7f94d6672a18 tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.206s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.529304] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.529650] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.529755] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.529944] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.530146] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.532366] env[61936]: INFO nova.compute.manager [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Terminating instance [ 1210.036517] env[61936]: DEBUG nova.compute.manager [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1210.036792] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1210.037724] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7677515-1abb-4097-b1ee-d9ce6f64b51d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.045853] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1210.046121] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77aa1b93-c417-4394-a9c3-5828a017f91d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.051965] env[61936]: DEBUG oslo_vmware.api [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1210.051965] env[61936]: value = "task-1253680" [ 1210.051965] env[61936]: _type = "Task" [ 1210.051965] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.059588] env[61936]: DEBUG oslo_vmware.api [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253680, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.561614] env[61936]: DEBUG oslo_vmware.api [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253680, 'name': PowerOffVM_Task, 'duration_secs': 0.168127} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.561990] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1210.562074] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1210.562302] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65115124-94c2-47d5-bd44-798fac8a50f8 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.624576] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1210.624811] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1210.624975] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Deleting the datastore file [datastore2] 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1210.625268] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1521132b-2ef0-4489-92b0-ced18d3195b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.631827] env[61936]: DEBUG oslo_vmware.api [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for the task: (returnval){ [ 1210.631827] env[61936]: value = "task-1253682" [ 1210.631827] env[61936]: _type = "Task" [ 1210.631827] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.639091] env[61936]: DEBUG oslo_vmware.api [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253682, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.141742] env[61936]: DEBUG oslo_vmware.api [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Task: {'id': task-1253682, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127356} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.141941] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1211.142144] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1211.142326] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1211.142500] env[61936]: INFO nova.compute.manager [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1211.142740] env[61936]: DEBUG oslo.service.loopingcall [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1211.142937] env[61936]: DEBUG nova.compute.manager [-] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1211.143076] env[61936]: DEBUG nova.network.neutron [-] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1211.658123] env[61936]: DEBUG nova.compute.manager [req-441a24a7-9aed-471d-a78f-2a69954527c3 req-5611bfc4-066c-4f88-b98d-079d9f5c7880 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Received event network-vif-deleted-2d477dba-b68d-487f-89a8-216750d10485 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1211.658436] env[61936]: INFO nova.compute.manager [req-441a24a7-9aed-471d-a78f-2a69954527c3 req-5611bfc4-066c-4f88-b98d-079d9f5c7880 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Neutron deleted interface 2d477dba-b68d-487f-89a8-216750d10485; detaching it from the instance and deleting it from the info cache [ 1211.658692] env[61936]: DEBUG nova.network.neutron [req-441a24a7-9aed-471d-a78f-2a69954527c3 req-5611bfc4-066c-4f88-b98d-079d9f5c7880 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.120430] env[61936]: DEBUG nova.network.neutron [-] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.164031] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-201f75dc-c10b-4527-ae8c-976a9376c64f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.172960] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cb007d-b8f6-4970-95eb-f935fdfeb648 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.199107] env[61936]: DEBUG nova.compute.manager [req-441a24a7-9aed-471d-a78f-2a69954527c3 req-5611bfc4-066c-4f88-b98d-079d9f5c7880 service nova] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Detach interface failed, port_id=2d477dba-b68d-487f-89a8-216750d10485, reason: Instance 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1212.624138] env[61936]: INFO nova.compute.manager [-] [instance: 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496] Took 1.48 seconds to deallocate network for instance. [ 1213.130441] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.130794] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.131061] env[61936]: DEBUG nova.objects.instance [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lazy-loading 'resources' on Instance uuid 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.696118] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddc4df1-94d6-4350-8180-835ee1c01012 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.703881] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466127ba-5546-4023-acf5-7d74371b37f2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.734091] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587921f8-796b-4137-8377-257004b7b61d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.740714] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982d23d4-555d-43fa-ba7b-f926f5f9f9ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.753318] env[61936]: DEBUG nova.compute.provider_tree [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1214.284543] env[61936]: DEBUG nova.scheduler.client.report [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Updated inventory for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with generation 138 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1214.284939] env[61936]: DEBUG nova.compute.provider_tree [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Updating resource provider abd382a7-b427-4dcc-a050-3fb56fdd60ac generation from 138 to 139 during operation: update_inventory {{(pid=61936) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1214.285040] env[61936]: DEBUG nova.compute.provider_tree [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Updating inventory in ProviderTree for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1214.789666] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.809306] env[61936]: INFO nova.scheduler.client.report [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Deleted allocations for instance 2c20a6cb-8d9b-4d4a-9c5e-3a3774292496 [ 1215.316597] env[61936]: DEBUG oslo_concurrency.lockutils [None req-eb02397f-d515-4489-a4d2-05aa44fa036d tempest-AttachVolumeTestJSON-1750295842 tempest-AttachVolumeTestJSON-1750295842-project-member] Lock "2c20a6cb-8d9b-4d4a-9c5e-3a3774292496" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.787s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.880902] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "2dd14535-9112-4464-a2d2-c1e139f3b132" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.881207] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.384141] env[61936]: DEBUG nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Starting instance... {{(pid=61936) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1218.907614] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.907904] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.909758] env[61936]: INFO nova.compute.claims [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1219.975782] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c443a741-146f-4e06-9584-c68e1d34563f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.983265] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b4c35d-39ab-4013-afbd-e4ad3014d8b2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.011682] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ef70c9-2261-4667-9f0a-d158d9e4242c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.018079] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98308ca9-0aef-47ea-8446-1af8206b5f93 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.030363] env[61936]: DEBUG nova.compute.provider_tree [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1220.533351] env[61936]: DEBUG nova.scheduler.client.report [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1221.039193] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.131s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1221.039555] env[61936]: DEBUG nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Start building networks asynchronously for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1221.544573] env[61936]: DEBUG nova.compute.utils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Using /dev/sd instead of None {{(pid=61936) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1221.545980] env[61936]: DEBUG nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Allocating IP information in the background. {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1221.547040] env[61936]: DEBUG nova.network.neutron [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] allocate_for_instance() {{(pid=61936) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1221.591643] env[61936]: DEBUG nova.policy [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f378ef8b8b5f4ae59c0d71dd1661bb59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa564b684410493fa0028fd345048e02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61936) authorize /opt/stack/nova/nova/policy.py:201}} [ 1221.853895] env[61936]: DEBUG nova.network.neutron [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Successfully created port: c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1222.049913] env[61936]: DEBUG nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Start building block device mappings for instance. {{(pid=61936) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1222.557711] env[61936]: INFO nova.virt.block_device [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Booting with volume b2b42df6-1e52-49bd-96f2-356efe929056 at /dev/sda [ 1222.590910] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ed5c39f-1ae7-4f7c-9621-70dd3f9ff6f9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.601668] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f52d4b-a2c0-481f-b92e-5e9e444267b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.628091] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70d92995-d2f1-4c35-8f42-6d68116aa4d9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.636435] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d017225-bcec-487f-93d6-95744177d20b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.662210] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0675b3d4-914f-4518-848f-8d2ab8557bd0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.668940] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5076fdef-1046-4c13-9d0e-d353d8fe042c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.682292] env[61936]: DEBUG nova.virt.block_device [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating existing volume attachment record: 9e5a43da-d768-4c21-a3b3-2000614571bb {{(pid=61936) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1223.227550] env[61936]: DEBUG nova.compute.manager [req-14683508-a3df-46cf-89a9-fefb91dc728e req-2b9e624a-129c-4ed1-a08f-b8c29710b881 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Received event network-vif-plugged-c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1223.227990] env[61936]: DEBUG oslo_concurrency.lockutils [req-14683508-a3df-46cf-89a9-fefb91dc728e req-2b9e624a-129c-4ed1-a08f-b8c29710b881 service nova] Acquiring lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.228210] env[61936]: DEBUG oslo_concurrency.lockutils [req-14683508-a3df-46cf-89a9-fefb91dc728e req-2b9e624a-129c-4ed1-a08f-b8c29710b881 service nova] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.228454] env[61936]: DEBUG oslo_concurrency.lockutils [req-14683508-a3df-46cf-89a9-fefb91dc728e req-2b9e624a-129c-4ed1-a08f-b8c29710b881 service nova] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.228664] env[61936]: DEBUG nova.compute.manager [req-14683508-a3df-46cf-89a9-fefb91dc728e req-2b9e624a-129c-4ed1-a08f-b8c29710b881 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] No waiting events found dispatching network-vif-plugged-c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1223.228873] env[61936]: WARNING nova.compute.manager [req-14683508-a3df-46cf-89a9-fefb91dc728e req-2b9e624a-129c-4ed1-a08f-b8c29710b881 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Received unexpected event network-vif-plugged-c2ad3fe4-ccd8-4edf-96a2-34446e36905d for instance with vm_state building and task_state block_device_mapping. [ 1223.310079] env[61936]: DEBUG nova.network.neutron [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Successfully updated port: c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1223.812923] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1223.813120] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1223.813265] env[61936]: DEBUG nova.network.neutron [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1224.343142] env[61936]: DEBUG nova.network.neutron [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Instance cache missing network info. {{(pid=61936) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1224.461655] env[61936]: DEBUG nova.network.neutron [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [{"id": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "address": "fa:16:3e:d4:59:1a", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ad3fe4-cc", "ovs_interfaceid": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.770028] env[61936]: DEBUG nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Start spawning the instance on the hypervisor. {{(pid=61936) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1224.770494] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1224.770706] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1224.770860] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1224.771054] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1224.771229] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1224.771364] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1224.771560] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1224.771715] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1224.771878] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1224.772049] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1224.772226] env[61936]: DEBUG nova.virt.hardware [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1224.773113] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498b485c-09f8-41d4-ac24-0f8402ff20b4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.781593] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dffe9c2-d1e8-4395-8609-cefd170f63b1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.964428] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1224.964764] env[61936]: DEBUG nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Instance network_info: |[{"id": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "address": "fa:16:3e:d4:59:1a", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ad3fe4-cc", "ovs_interfaceid": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61936) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1224.965221] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:59:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2ad3fe4-ccd8-4edf-96a2-34446e36905d', 'vif_model': 'vmxnet3'}] {{(pid=61936) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1224.973542] env[61936]: DEBUG oslo.service.loopingcall [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1224.973751] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Creating VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1224.973998] env[61936]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7e2bc01-b170-4f05-acd8-3fd7b471a36e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.993675] env[61936]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1224.993675] env[61936]: value = "task-1253688" [ 1224.993675] env[61936]: _type = "Task" [ 1224.993675] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.000504] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253688, 'name': CreateVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.251435] env[61936]: DEBUG nova.compute.manager [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Received event network-changed-c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1225.251617] env[61936]: DEBUG nova.compute.manager [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Refreshing instance network info cache due to event network-changed-c2ad3fe4-ccd8-4edf-96a2-34446e36905d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1225.251838] env[61936]: DEBUG oslo_concurrency.lockutils [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] Acquiring lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1225.251987] env[61936]: DEBUG oslo_concurrency.lockutils [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] Acquired lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.252165] env[61936]: DEBUG nova.network.neutron [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Refreshing network info cache for port c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1225.503905] env[61936]: DEBUG oslo_vmware.api [-] Task: {'id': task-1253688, 'name': CreateVM_Task, 'duration_secs': 0.301756} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.504262] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Created VM on the ESX host {{(pid=61936) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1225.504769] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270058', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'name': 'volume-b2b42df6-1e52-49bd-96f2-356efe929056', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2dd14535-9112-4464-a2d2-c1e139f3b132', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'serial': 'b2b42df6-1e52-49bd-96f2-356efe929056'}, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'disk_bus': None, 'device_type': None, 'attachment_id': '9e5a43da-d768-4c21-a3b3-2000614571bb', 'volume_type': None}], 'swap': None} {{(pid=61936) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1225.504986] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Root volume attach. Driver type: vmdk {{(pid=61936) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1225.505771] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bf5ab5-c320-4c47-b7c5-2a25282d5ee3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.513094] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9156469e-93a3-4a1f-8577-58d5416b85e1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.518662] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccd07ca-eaea-42c6-abde-fbf4a16d9f3f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.524199] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-408a0cf3-8f70-4c86-9bf3-7411191f4c6c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.530011] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1225.530011] env[61936]: value = "task-1253689" [ 1225.530011] env[61936]: _type = "Task" [ 1225.530011] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.540633] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253689, 'name': RelocateVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.040530] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253689, 'name': RelocateVM_Task, 'duration_secs': 0.026213} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.040820] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Volume attach. Driver type: vmdk {{(pid=61936) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1226.042187] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270058', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'name': 'volume-b2b42df6-1e52-49bd-96f2-356efe929056', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2dd14535-9112-4464-a2d2-c1e139f3b132', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'serial': 'b2b42df6-1e52-49bd-96f2-356efe929056'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1226.042976] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3223d65-797d-4ca5-9c19-fba0b2ca55f5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.063152] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5b05a3-3911-47ba-863f-d3ac80c1852e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.087099] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-b2b42df6-1e52-49bd-96f2-356efe929056/volume-b2b42df6-1e52-49bd-96f2-356efe929056.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1226.089648] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60712cf5-5427-40eb-87c3-d46d2550e4a4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.109474] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1226.109474] env[61936]: value = "task-1253690" [ 1226.109474] env[61936]: _type = "Task" [ 1226.109474] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.119875] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253690, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.164153] env[61936]: DEBUG nova.network.neutron [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updated VIF entry in instance network info cache for port c2ad3fe4-ccd8-4edf-96a2-34446e36905d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1226.164721] env[61936]: DEBUG nova.network.neutron [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [{"id": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "address": "fa:16:3e:d4:59:1a", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ad3fe4-cc", "ovs_interfaceid": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.620090] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253690, 'name': ReconfigVM_Task, 'duration_secs': 0.248236} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.620463] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-b2b42df6-1e52-49bd-96f2-356efe929056/volume-b2b42df6-1e52-49bd-96f2-356efe929056.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1226.625748] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10925c80-e018-47e4-ab93-5754d1fea4de {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.641754] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1226.641754] env[61936]: value = "task-1253691" [ 1226.641754] env[61936]: _type = "Task" [ 1226.641754] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.649986] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.667829] env[61936]: DEBUG oslo_concurrency.lockutils [req-407794b1-8727-4fea-b034-0c8575eb0532 req-cf4cc980-f388-43f3-a557-2a86569b79b8 service nova] Releasing lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.152715] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253691, 'name': ReconfigVM_Task, 'duration_secs': 0.120951} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.153023] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270058', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'name': 'volume-b2b42df6-1e52-49bd-96f2-356efe929056', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2dd14535-9112-4464-a2d2-c1e139f3b132', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'serial': 'b2b42df6-1e52-49bd-96f2-356efe929056'} {{(pid=61936) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1227.153579] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79f9cebf-6f52-486f-b1e0-2f669e5de414 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.159602] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1227.159602] env[61936]: value = "task-1253692" [ 1227.159602] env[61936]: _type = "Task" [ 1227.159602] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.167421] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253692, 'name': Rename_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.670423] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253692, 'name': Rename_Task, 'duration_secs': 0.119948} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.670767] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1227.670899] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb78d490-2677-4b46-a4bf-4998d6ffadcb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.677022] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1227.677022] env[61936]: value = "task-1253693" [ 1227.677022] env[61936]: _type = "Task" [ 1227.677022] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.684021] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.186112] env[61936]: DEBUG oslo_vmware.api [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253693, 'name': PowerOnVM_Task, 'duration_secs': 0.484938} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.186322] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1228.186517] env[61936]: INFO nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Took 3.42 seconds to spawn the instance on the hypervisor. [ 1228.186704] env[61936]: DEBUG nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Checking state {{(pid=61936) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1228.187489] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92cd7fa-1409-4ed5-85e1-715cc7e12657 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.705889] env[61936]: INFO nova.compute.manager [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Took 9.82 seconds to build instance. [ 1229.208569] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8d8dc18b-2d9a-4c6a-845a-fe78279b5818 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.327s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.049124] env[61936]: DEBUG nova.compute.manager [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Received event network-changed-855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1230.049376] env[61936]: DEBUG nova.compute.manager [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Refreshing instance network info cache due to event network-changed-855dd2af-6f16-4185-88e2-0022244dc30d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1230.049623] env[61936]: DEBUG oslo_concurrency.lockutils [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] Acquiring lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.049777] env[61936]: DEBUG oslo_concurrency.lockutils [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] Acquired lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.049953] env[61936]: DEBUG nova.network.neutron [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Refreshing network info cache for port 855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1230.767801] env[61936]: DEBUG nova.network.neutron [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updated VIF entry in instance network info cache for port 855dd2af-6f16-4185-88e2-0022244dc30d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1230.768253] env[61936]: DEBUG nova.network.neutron [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updating instance_info_cache with network_info: [{"id": "855dd2af-6f16-4185-88e2-0022244dc30d", "address": "fa:16:3e:b1:24:f2", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855dd2af-6f", "ovs_interfaceid": "855dd2af-6f16-4185-88e2-0022244dc30d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.271770] env[61936]: DEBUG oslo_concurrency.lockutils [req-85bf0fc2-c4c1-4ed3-aa91-7255af724ad6 req-ef6b590c-d901-47ed-8165-9af6561fc3ef service nova] Releasing lock "refresh_cache-e6930eaf-bb63-4553-8cd2-f2eb62fca510" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1231.444694] env[61936]: DEBUG nova.compute.manager [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Stashing vm_state: active {{(pid=61936) _prep_resize /opt/stack/nova/nova/compute/manager.py:6015}} [ 1231.963011] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.963329] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.074210] env[61936]: DEBUG nova.compute.manager [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Received event network-changed-c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1232.074474] env[61936]: DEBUG nova.compute.manager [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Refreshing instance network info cache due to event network-changed-c2ad3fe4-ccd8-4edf-96a2-34446e36905d. {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11534}} [ 1232.074731] env[61936]: DEBUG oslo_concurrency.lockutils [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] Acquiring lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1232.074925] env[61936]: DEBUG oslo_concurrency.lockutils [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] Acquired lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1232.075140] env[61936]: DEBUG nova.network.neutron [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Refreshing network info cache for port c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1232.468547] env[61936]: INFO nova.compute.claims [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1232.777277] env[61936]: DEBUG nova.network.neutron [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updated VIF entry in instance network info cache for port c2ad3fe4-ccd8-4edf-96a2-34446e36905d. {{(pid=61936) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1232.777636] env[61936]: DEBUG nova.network.neutron [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [{"id": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "address": "fa:16:3e:d4:59:1a", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ad3fe4-cc", "ovs_interfaceid": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.974563] env[61936]: INFO nova.compute.resource_tracker [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating resource usage from migration e71445ba-30c4-4d44-ba4e-de64edc8b3f2 [ 1233.041498] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88af7286-1745-42be-aaa6-9414603f78af {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.052049] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cb0506-3394-4125-8af4-9eb5f6d5a8a9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.082733] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2db6e56-00f2-4d94-aac4-8f8e986662ea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.089352] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f99c26-d18e-444b-939e-e00d28b6cea2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.101653] env[61936]: DEBUG nova.compute.provider_tree [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.280847] env[61936]: DEBUG oslo_concurrency.lockutils [req-27c64083-7861-4ca4-bbbb-e121f3ff1d4c req-0a67f851-c35d-498f-98cc-b788ec32d374 service nova] Releasing lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1233.605163] env[61936]: DEBUG nova.scheduler.client.report [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1234.110768] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.147s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.110990] env[61936]: INFO nova.compute.manager [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Migrating [ 1234.627498] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.627885] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.627885] env[61936]: DEBUG nova.network.neutron [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1235.322411] env[61936]: DEBUG nova.network.neutron [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [{"id": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "address": "fa:16:3e:d4:59:1a", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ad3fe4-cc", "ovs_interfaceid": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.825701] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.341010] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267d5496-bb99-41f3-a5db-8cd593426040 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.359291] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance '2dd14535-9112-4464-a2d2-c1e139f3b132' progress to 0 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1237.864825] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1237.865172] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-582f5687-b5b4-4bbf-843a-41e04fe7b0c3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.872648] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1237.872648] env[61936]: value = "task-1253694" [ 1237.872648] env[61936]: _type = "Task" [ 1237.872648] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.880169] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253694, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.382921] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253694, 'name': PowerOffVM_Task, 'duration_secs': 0.213229} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.383297] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1238.383417] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance '2dd14535-9112-4464-a2d2-c1e139f3b132' progress to 17 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1238.889562] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1238.889861] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1238.890015] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1238.890209] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1238.890364] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1238.890555] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1238.890771] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1238.890935] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1238.891119] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1238.891303] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1238.891522] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1238.896798] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a77091f-29a3-45d9-96f2-814a7d910272 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.912603] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1238.912603] env[61936]: value = "task-1253695" [ 1238.912603] env[61936]: _type = "Task" [ 1238.912603] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.920325] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253695, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.422641] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253695, 'name': ReconfigVM_Task, 'duration_secs': 0.165438} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.423039] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance '2dd14535-9112-4464-a2d2-c1e139f3b132' progress to 33 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1239.930976] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T16:42:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1239.930976] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1239.930976] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image limits 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1239.930976] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Flavor pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1239.931255] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Image pref 0:0:0 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1239.931291] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61936) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1239.931479] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1239.931633] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1239.931799] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Got 1 possible topologies {{(pid=61936) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1239.931961] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1239.932197] env[61936]: DEBUG nova.virt.hardware [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61936) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1239.938383] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1239.938383] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a0b1a24-4fb5-42c3-852c-281f0a8684aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.956020] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1239.956020] env[61936]: value = "task-1253696" [ 1239.956020] env[61936]: _type = "Task" [ 1239.956020] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.963754] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253696, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.324201] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.324492] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.324726] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.324920] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.325116] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.327208] env[61936]: INFO nova.compute.manager [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Terminating instance [ 1240.455472] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.465893] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253696, 'name': ReconfigVM_Task, 'duration_secs': 0.142159} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.466168] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1240.466897] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3903d1fa-b289-4909-811b-be732bde9f3b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.487797] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-b2b42df6-1e52-49bd-96f2-356efe929056/volume-b2b42df6-1e52-49bd-96f2-356efe929056.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1240.488027] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53300a30-dc2d-4a22-9d0a-1cca2760a52b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.504803] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1240.504803] env[61936]: value = "task-1253697" [ 1240.504803] env[61936]: _type = "Task" [ 1240.504803] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.511897] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253697, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.830695] env[61936]: DEBUG nova.compute.manager [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1240.830948] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1240.831882] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea404226-2016-4547-8890-432681aeb126 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.839197] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1240.839427] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38efb6fd-2030-4efe-a906-d813abafa752 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.844733] env[61936]: DEBUG oslo_vmware.api [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1240.844733] env[61936]: value = "task-1253698" [ 1240.844733] env[61936]: _type = "Task" [ 1240.844733] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.854055] env[61936]: DEBUG oslo_vmware.api [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.014384] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253697, 'name': ReconfigVM_Task, 'duration_secs': 0.250138} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.014705] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-b2b42df6-1e52-49bd-96f2-356efe929056/volume-b2b42df6-1e52-49bd-96f2-356efe929056.vmdk or device None with type thin {{(pid=61936) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1241.014984] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance '2dd14535-9112-4464-a2d2-c1e139f3b132' progress to 50 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1241.354579] env[61936]: DEBUG oslo_vmware.api [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253698, 'name': PowerOffVM_Task, 'duration_secs': 0.188171} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.354883] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1241.355075] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1241.355331] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae3df360-ea0e-488d-8a77-fdaac76c1c72 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.417279] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1241.417506] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1241.417698] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleting the datastore file [datastore1] 767a9a43-784e-4270-92c7-53b55122dd8f {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1241.417973] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-514f1b6f-828b-4ec0-9b86-8faf0728cceb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.424462] env[61936]: DEBUG oslo_vmware.api [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for the task: (returnval){ [ 1241.424462] env[61936]: value = "task-1253700" [ 1241.424462] env[61936]: _type = "Task" [ 1241.424462] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.431715] env[61936]: DEBUG oslo_vmware.api [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.455285] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1241.455450] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 1241.521236] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b03ca26-bfcf-4490-8ea8-fc607775a45e {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.541315] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde6f1c6-015b-4c86-ba4d-c7d0f75cd3d1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.559321] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance '2dd14535-9112-4464-a2d2-c1e139f3b132' progress to 67 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1241.938692] env[61936]: DEBUG oslo_vmware.api [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Task: {'id': task-1253700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11931} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.939123] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1241.939429] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1241.939730] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1241.940028] env[61936]: INFO nova.compute.manager [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1241.940407] env[61936]: DEBUG oslo.service.loopingcall [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1241.940700] env[61936]: DEBUG nova.compute.manager [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1241.940858] env[61936]: DEBUG nova.network.neutron [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1242.395443] env[61936]: DEBUG nova.compute.manager [req-8f383e94-a0a8-4426-b150-62e136c8f4a9 req-d661449a-2667-4f63-a37e-9f1cdd7b7157 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Received event network-vif-deleted-77480d6a-c7f5-41bf-9dc1-6554e54c4552 {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1242.395443] env[61936]: INFO nova.compute.manager [req-8f383e94-a0a8-4426-b150-62e136c8f4a9 req-d661449a-2667-4f63-a37e-9f1cdd7b7157 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Neutron deleted interface 77480d6a-c7f5-41bf-9dc1-6554e54c4552; detaching it from the instance and deleting it from the info cache [ 1242.395443] env[61936]: DEBUG nova.network.neutron [req-8f383e94-a0a8-4426-b150-62e136c8f4a9 req-d661449a-2667-4f63-a37e-9f1cdd7b7157 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.449902] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.454537] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.454707] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 1242.877921] env[61936]: DEBUG nova.network.neutron [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.898152] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3392802-8c58-4cde-82c5-ca42cc779cf6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.907122] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f938730b-0a1a-46fb-b55b-5bef37550beb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.932844] env[61936]: DEBUG nova.compute.manager [req-8f383e94-a0a8-4426-b150-62e136c8f4a9 req-d661449a-2667-4f63-a37e-9f1cdd7b7157 service nova] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Detach interface failed, port_id=77480d6a-c7f5-41bf-9dc1-6554e54c4552, reason: Instance 767a9a43-784e-4270-92c7-53b55122dd8f could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1242.957095] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Didn't find any instances for network info cache update. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10412}} [ 1242.957307] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1242.957486] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1243.232442] env[61936]: DEBUG nova.network.neutron [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Port c2ad3fe4-ccd8-4edf-96a2-34446e36905d binding to destination host cpu-1 is already ACTIVE {{(pid=61936) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1243.380635] env[61936]: INFO nova.compute.manager [-] [instance: 767a9a43-784e-4270-92c7-53b55122dd8f] Took 1.44 seconds to deallocate network for instance. [ 1243.460394] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.460655] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.460847] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.461012] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1243.461964] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48985372-7600-40c5-9f35-c34c48cf1498 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.470576] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae160a4a-7cdd-44e8-8656-b5b0ccb841bb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.485404] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261afd0c-d254-4f66-ae41-72a695293cea {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.491394] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05d5778-6aee-4102-9e72-3e65ec8fc957 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.518901] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180773MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1243.519058] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.519256] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.887508] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.253837] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.254048] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.254237] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.526938] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Applying migration context for instance 2dd14535-9112-4464-a2d2-c1e139f3b132 as it has an incoming, in-progress migration e71445ba-30c4-4d44-ba4e-de64edc8b3f2. Migration status is post-migrating {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1244.527700] env[61936]: INFO nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating resource usage from migration e71445ba-30c4-4d44-ba4e-de64edc8b3f2 [ 1244.546294] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance e6930eaf-bb63-4553-8cd2-f2eb62fca510 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1244.546442] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 149ac6e4-1233-4fec-8310-1794f9273b93 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1244.546610] env[61936]: WARNING nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 767a9a43-784e-4270-92c7-53b55122dd8f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1244.546738] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Migration e71445ba-30c4-4d44-ba4e-de64edc8b3f2 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1244.546857] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Instance 2dd14535-9112-4464-a2d2-c1e139f3b132 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61936) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1244.547041] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1244.547185] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1244.607021] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634a806b-6301-4b7f-8dfd-14bf1ef37e16 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.614333] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab574fbd-72b5-4f42-8251-fe43e77cd5f6 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.646768] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89777ee5-6afe-4d76-a53e-10422506881c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.653469] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa087f64-9c70-446d-9ece-cc783ca39be1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.665971] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1245.168642] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1245.288840] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1245.289066] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1245.289265] env[61936]: DEBUG nova.network.neutron [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1245.673843] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1245.674073] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.155s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.674386] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.787s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.674583] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.691999] env[61936]: INFO nova.scheduler.client.report [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Deleted allocations for instance 767a9a43-784e-4270-92c7-53b55122dd8f [ 1246.048716] env[61936]: DEBUG nova.network.neutron [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [{"id": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "address": "fa:16:3e:d4:59:1a", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ad3fe4-cc", "ovs_interfaceid": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1246.174655] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.175027] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.199747] env[61936]: DEBUG oslo_concurrency.lockutils [None req-31d3b216-9bab-4416-ad8d-62b05faa2aed tempest-AttachVolumeShelveTestJSON-2109134238 tempest-AttachVolumeShelveTestJSON-2109134238-project-member] Lock "767a9a43-784e-4270-92c7-53b55122dd8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.875s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1246.551228] env[61936]: DEBUG oslo_concurrency.lockutils [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1247.061314] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b83cb1e-b866-4887-a414-0bc83fe21a6b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.068840] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5cf612-91f8-4849-829e-7b35f68df951 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.161151] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe16db82-ab54-4464-8d92-8a9d50ff1b52 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.179523] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2195f3c1-ed70-4c74-9fa5-8b35af77d6c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.185817] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance '2dd14535-9112-4464-a2d2-c1e139f3b132' progress to 83 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1248.450158] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.693261] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powering on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1248.693610] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd01105b-df8e-4196-9d50-867b1408573a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.700727] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1248.700727] env[61936]: value = "task-1253702" [ 1248.700727] env[61936]: _type = "Task" [ 1248.700727] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.716961] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253702, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.954947] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.211169] env[61936]: DEBUG oslo_vmware.api [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253702, 'name': PowerOnVM_Task, 'duration_secs': 0.360172} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.211522] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powered on the VM {{(pid=61936) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1249.212304] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-64420452-6f7a-4856-a097-a1618582aaef tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance '2dd14535-9112-4464-a2d2-c1e139f3b132' progress to 100 {{(pid=61936) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1252.065865] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "2dd14535-9112-4464-a2d2-c1e139f3b132" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.066695] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.066695] env[61936]: DEBUG nova.compute.manager [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Going to confirm migration 4 {{(pid=61936) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5174}} [ 1252.626122] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1252.626122] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquired lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.626122] env[61936]: DEBUG nova.network.neutron [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Building network info cache for instance {{(pid=61936) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1252.626363] env[61936]: DEBUG nova.objects.instance [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'info_cache' on Instance uuid 2dd14535-9112-4464-a2d2-c1e139f3b132 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.851312] env[61936]: DEBUG nova.network.neutron [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [{"id": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "address": "fa:16:3e:d4:59:1a", "network": {"id": "bcc1f7b2-1d99-4860-b769-9101ce9a1440", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-789384208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa564b684410493fa0028fd345048e02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2ad3fe4-cc", "ovs_interfaceid": "c2ad3fe4-ccd8-4edf-96a2-34446e36905d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1254.354348] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Releasing lock "refresh_cache-2dd14535-9112-4464-a2d2-c1e139f3b132" {{(pid=61936) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1254.354618] env[61936]: DEBUG nova.objects.instance [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'migration_context' on Instance uuid 2dd14535-9112-4464-a2d2-c1e139f3b132 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1254.857269] env[61936]: DEBUG nova.objects.base [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Object Instance<2dd14535-9112-4464-a2d2-c1e139f3b132> lazy-loaded attributes: info_cache,migration_context {{(pid=61936) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1254.858257] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e766fd-6cfc-4ef9-a6ae-93680b5190a5 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.877166] env[61936]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5287ee9-7d22-4120-abfe-13abfa50290d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.882703] env[61936]: DEBUG oslo_vmware.api [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1254.882703] env[61936]: value = "session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521b8810-e0d6-4320-de15-4b0e7a527af7" [ 1254.882703] env[61936]: _type = "Task" [ 1254.882703] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.890008] env[61936]: DEBUG oslo_vmware.api [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521b8810-e0d6-4320-de15-4b0e7a527af7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.393851] env[61936]: DEBUG oslo_vmware.api [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': session[52f24af7-fb5c-9bb1-47b3-cdb662ede748]521b8810-e0d6-4320-de15-4b0e7a527af7, 'name': SearchDatastore_Task, 'duration_secs': 0.010729} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.394187] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.394448] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.958336] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b54138-08d8-400f-9fea-446d47406c8f {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.966148] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d4a783-62db-467d-9649-57d2327d0fe1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.994731] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c45d16-03d0-4de6-8a1c-53313a5b5dd7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.001502] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efad251e-4b86-4e5c-8909-ecb672e22786 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.015393] env[61936]: DEBUG nova.compute.provider_tree [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1256.518356] env[61936]: DEBUG nova.scheduler.client.report [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1257.528303] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.134s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1258.083363] env[61936]: INFO nova.scheduler.client.report [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted allocation for migration e71445ba-30c4-4d44-ba4e-de64edc8b3f2 [ 1258.468608] env[61936]: INFO nova.compute.manager [None req-3930b773-2bfc-43c4-8af0-d4b7dedfcf8a tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Get console output [ 1258.468969] env[61936]: WARNING nova.virt.vmwareapi.driver [None req-3930b773-2bfc-43c4-8af0-d4b7dedfcf8a tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] The console log is missing. Check your VSPC configuration [ 1258.589017] env[61936]: DEBUG oslo_concurrency.lockutils [None req-e01d012f-ef27-4ea0-98b3-7e32909fe9ba tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.523s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.140657] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "2dd14535-9112-4464-a2d2-c1e139f3b132" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.141137] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.141186] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.141358] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.141533] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.145048] env[61936]: INFO nova.compute.manager [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Terminating instance [ 1285.649495] env[61936]: DEBUG nova.compute.manager [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1285.649740] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1285.650115] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cc338de-95f1-46cf-9a0c-0817eb4227d9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.657727] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1285.657727] env[61936]: value = "task-1253703" [ 1285.657727] env[61936]: _type = "Task" [ 1285.657727] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.666625] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.167434] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253703, 'name': PowerOffVM_Task, 'duration_secs': 0.164337} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.167785] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1286.167882] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Volume detach. Driver type: vmdk {{(pid=61936) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1286.168086] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270058', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'name': 'volume-b2b42df6-1e52-49bd-96f2-356efe929056', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '2dd14535-9112-4464-a2d2-c1e139f3b132', 'attached_at': '2024-10-10T16:55:09.000000', 'detached_at': '', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'serial': 'b2b42df6-1e52-49bd-96f2-356efe929056'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1286.168834] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bbd8dd-9c8d-4408-b3db-d9276232e285 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.186425] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e2da40-84be-44c2-b402-edf87b3cf510 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.192293] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1b2af9-22a9-4b31-a50d-a67bda0f8e0b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.208921] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33606c3b-83e6-40a7-ab19-55aab4a947f7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.222784] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] The volume has not been displaced from its original location: [datastore1] volume-b2b42df6-1e52-49bd-96f2-356efe929056/volume-b2b42df6-1e52-49bd-96f2-356efe929056.vmdk. No consolidation needed. {{(pid=61936) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1286.228008] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1286.228283] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8fb54a6b-164f-4e9a-8261-076320262d43 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.244691] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1286.244691] env[61936]: value = "task-1253704" [ 1286.244691] env[61936]: _type = "Task" [ 1286.244691] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.251937] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253704, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.754071] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253704, 'name': ReconfigVM_Task, 'duration_secs': 0.145903} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.754377] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=61936) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1286.758948] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13277fbd-fb5e-4032-8b1b-c02d17069505 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.775879] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1286.775879] env[61936]: value = "task-1253705" [ 1286.775879] env[61936]: _type = "Task" [ 1286.775879] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.786132] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.285594] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253705, 'name': ReconfigVM_Task, 'duration_secs': 0.119264} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.285963] env[61936]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-270058', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'name': 'volume-b2b42df6-1e52-49bd-96f2-356efe929056', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '2dd14535-9112-4464-a2d2-c1e139f3b132', 'attached_at': '2024-10-10T16:55:09.000000', 'detached_at': '', 'volume_id': 'b2b42df6-1e52-49bd-96f2-356efe929056', 'serial': 'b2b42df6-1e52-49bd-96f2-356efe929056'} {{(pid=61936) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1287.286192] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1287.286935] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26cf927c-f85e-4151-b477-384e92727eb1 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.293152] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1287.293367] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a07627ac-b40c-42c8-8238-7cf89221065a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.355858] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1287.356093] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1287.356280] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleting the datastore file [datastore1] 2dd14535-9112-4464-a2d2-c1e139f3b132 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1287.356538] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe399ac2-6db9-4d5a-97f9-e61adb913965 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.361954] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1287.361954] env[61936]: value = "task-1253707" [ 1287.361954] env[61936]: _type = "Task" [ 1287.361954] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.369492] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.873603] env[61936]: DEBUG oslo_vmware.api [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.077985} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.873852] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1287.874088] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1287.874262] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1287.874436] env[61936]: INFO nova.compute.manager [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1287.874676] env[61936]: DEBUG oslo.service.loopingcall [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1287.874873] env[61936]: DEBUG nova.compute.manager [-] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1287.874966] env[61936]: DEBUG nova.network.neutron [-] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1288.334943] env[61936]: DEBUG nova.compute.manager [req-3ce059f7-6c68-4584-bc71-91b72b154f6f req-fa08bd08-a003-4fa6-95bb-3a07b346356b service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Received event network-vif-deleted-c2ad3fe4-ccd8-4edf-96a2-34446e36905d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1288.335208] env[61936]: INFO nova.compute.manager [req-3ce059f7-6c68-4584-bc71-91b72b154f6f req-fa08bd08-a003-4fa6-95bb-3a07b346356b service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Neutron deleted interface c2ad3fe4-ccd8-4edf-96a2-34446e36905d; detaching it from the instance and deleting it from the info cache [ 1288.335365] env[61936]: DEBUG nova.network.neutron [req-3ce059f7-6c68-4584-bc71-91b72b154f6f req-fa08bd08-a003-4fa6-95bb-3a07b346356b service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.816543] env[61936]: DEBUG nova.network.neutron [-] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.837587] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aece5212-b06b-4818-a748-9d94361bbab4 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.847635] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284932d7-75a2-461f-b0d1-d11e9e89e3eb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.870094] env[61936]: DEBUG nova.compute.manager [req-3ce059f7-6c68-4584-bc71-91b72b154f6f req-fa08bd08-a003-4fa6-95bb-3a07b346356b service nova] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Detach interface failed, port_id=c2ad3fe4-ccd8-4edf-96a2-34446e36905d, reason: Instance 2dd14535-9112-4464-a2d2-c1e139f3b132 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1289.318837] env[61936]: INFO nova.compute.manager [-] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Took 1.44 seconds to deallocate network for instance. [ 1289.864402] env[61936]: INFO nova.compute.manager [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Took 0.55 seconds to detach 1 volumes for instance. [ 1289.867393] env[61936]: DEBUG nova.compute.manager [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 2dd14535-9112-4464-a2d2-c1e139f3b132] Deleting volume: b2b42df6-1e52-49bd-96f2-356efe929056 {{(pid=61936) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 1290.424954] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1290.425205] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1290.425400] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1290.448589] env[61936]: INFO nova.scheduler.client.report [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted allocations for instance 2dd14535-9112-4464-a2d2-c1e139f3b132 [ 1290.956077] env[61936]: DEBUG oslo_concurrency.lockutils [None req-ebc646f9-f030-4e4e-a268-df4a4d76d14f tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "2dd14535-9112-4464-a2d2-c1e139f3b132" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.815s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.972195] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "149ac6e4-1233-4fec-8310-1794f9273b93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1291.972556] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "149ac6e4-1233-4fec-8310-1794f9273b93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1291.972741] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "149ac6e4-1233-4fec-8310-1794f9273b93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1291.972949] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "149ac6e4-1233-4fec-8310-1794f9273b93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1291.973155] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "149ac6e4-1233-4fec-8310-1794f9273b93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.975513] env[61936]: INFO nova.compute.manager [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Terminating instance [ 1292.479649] env[61936]: DEBUG nova.compute.manager [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1292.479937] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1292.480845] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a83fc8-10ff-4fb5-87bc-570990fb9b93 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.489731] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1292.489955] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c85d836b-f7ef-4c2a-9e44-de17a3939559 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.495810] env[61936]: DEBUG oslo_vmware.api [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1292.495810] env[61936]: value = "task-1253709" [ 1292.495810] env[61936]: _type = "Task" [ 1292.495810] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.503757] env[61936]: DEBUG oslo_vmware.api [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253709, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.005256] env[61936]: DEBUG oslo_vmware.api [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253709, 'name': PowerOffVM_Task, 'duration_secs': 0.180711} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.005669] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1293.005775] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1293.005949] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6970fc7-4769-43ce-91ad-397a03998913 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.069875] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1293.070124] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Deleting contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1293.070315] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleting the datastore file [datastore1] 149ac6e4-1233-4fec-8310-1794f9273b93 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1293.070582] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b29ee8f5-baf8-43d5-84d7-c808b52bf09b {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.076509] env[61936]: DEBUG oslo_vmware.api [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1293.076509] env[61936]: value = "task-1253711" [ 1293.076509] env[61936]: _type = "Task" [ 1293.076509] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.083516] env[61936]: DEBUG oslo_vmware.api [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253711, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.586291] env[61936]: DEBUG oslo_vmware.api [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253711, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144831} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.586589] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1293.586787] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Deleted contents of the VM from datastore datastore1 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1293.586972] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1293.587171] env[61936]: INFO nova.compute.manager [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1293.587416] env[61936]: DEBUG oslo.service.loopingcall [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1293.587613] env[61936]: DEBUG nova.compute.manager [-] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1293.587709] env[61936]: DEBUG nova.network.neutron [-] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1293.830044] env[61936]: DEBUG nova.compute.manager [req-48855653-b6d0-4dbe-a9d9-b1cf8861d2dc req-04f85d91-67cd-4fa3-9204-88642ee68f67 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Received event network-vif-deleted-545930bc-8fe4-41d2-8128-911d03a7bbfe {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1293.830223] env[61936]: INFO nova.compute.manager [req-48855653-b6d0-4dbe-a9d9-b1cf8861d2dc req-04f85d91-67cd-4fa3-9204-88642ee68f67 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Neutron deleted interface 545930bc-8fe4-41d2-8128-911d03a7bbfe; detaching it from the instance and deleting it from the info cache [ 1293.830397] env[61936]: DEBUG nova.network.neutron [req-48855653-b6d0-4dbe-a9d9-b1cf8861d2dc req-04f85d91-67cd-4fa3-9204-88642ee68f67 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1294.309296] env[61936]: DEBUG nova.network.neutron [-] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1294.333259] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75887155-fd6f-4b48-af0e-4f1688fd1b88 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.343098] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0a2625-288e-4454-9181-539a62c660bf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.364754] env[61936]: DEBUG nova.compute.manager [req-48855653-b6d0-4dbe-a9d9-b1cf8861d2dc req-04f85d91-67cd-4fa3-9204-88642ee68f67 service nova] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Detach interface failed, port_id=545930bc-8fe4-41d2-8128-911d03a7bbfe, reason: Instance 149ac6e4-1233-4fec-8310-1794f9273b93 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1294.812488] env[61936]: INFO nova.compute.manager [-] [instance: 149ac6e4-1233-4fec-8310-1794f9273b93] Took 1.22 seconds to deallocate network for instance. [ 1295.318861] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.319288] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.319395] env[61936]: DEBUG nova.objects.instance [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'resources' on Instance uuid 149ac6e4-1233-4fec-8310-1794f9273b93 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1295.863907] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2b792d-47c7-4d8c-9320-209df4eb24aa {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.873254] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4669a16b-2f74-42cb-ac1c-8f1f2829bf6a {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.902634] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cfdac60-c10d-4ae6-b3be-f0ca60a42a76 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.909530] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9380a3a3-9c66-482d-aa24-b0f58a0cb1b3 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.922278] env[61936]: DEBUG nova.compute.provider_tree [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1296.425525] env[61936]: DEBUG nova.scheduler.client.report [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1296.930274] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.949165] env[61936]: INFO nova.scheduler.client.report [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted allocations for instance 149ac6e4-1233-4fec-8310-1794f9273b93 [ 1297.457326] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4dfed3e6-bbd5-41b1-abe9-80bc48583f72 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "149ac6e4-1233-4fec-8310-1794f9273b93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.485s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1300.456066] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1301.455583] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1301.455834] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61936) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10945}} [ 1301.650033] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1301.650395] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1301.650570] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1301.650786] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1301.650973] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.653390] env[61936]: INFO nova.compute.manager [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Terminating instance [ 1302.157632] env[61936]: DEBUG nova.compute.manager [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Start destroying the instance on the hypervisor. {{(pid=61936) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1302.157868] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Destroying instance {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1302.158768] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d92f01-2cf1-4831-bc8c-fe52048a9924 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.166669] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Powering off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1302.166908] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed40dba4-6221-43dd-a337-835fc9d8d36c {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.173314] env[61936]: DEBUG oslo_vmware.api [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1302.173314] env[61936]: value = "task-1253713" [ 1302.173314] env[61936]: _type = "Task" [ 1302.173314] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.181099] env[61936]: DEBUG oslo_vmware.api [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.682760] env[61936]: DEBUG oslo_vmware.api [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253713, 'name': PowerOffVM_Task, 'duration_secs': 0.207069} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.683199] env[61936]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Powered off the VM {{(pid=61936) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1302.683199] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Unregistering the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1302.683415] env[61936]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aaea9ec6-d086-43c1-b581-0abb521be593 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.747638] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Unregistered the VM {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1302.747882] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Deleting contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1302.748051] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleting the datastore file [datastore2] e6930eaf-bb63-4553-8cd2-f2eb62fca510 {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1302.748317] env[61936]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0de285ed-aef4-44a6-a888-10f5d2a5b518 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.755182] env[61936]: DEBUG oslo_vmware.api [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for the task: (returnval){ [ 1302.755182] env[61936]: value = "task-1253715" [ 1302.755182] env[61936]: _type = "Task" [ 1302.755182] env[61936]: } to complete. {{(pid=61936) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.763084] env[61936]: DEBUG oslo_vmware.api [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253715, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.265223] env[61936]: DEBUG oslo_vmware.api [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Task: {'id': task-1253715, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132624} completed successfully. {{(pid=61936) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.265539] env[61936]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted the datastore file {{(pid=61936) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1303.265774] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Deleted contents of the VM from datastore datastore2 {{(pid=61936) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1303.265992] env[61936]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Instance destroyed {{(pid=61936) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1303.266236] env[61936]: INFO nova.compute.manager [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1303.266520] env[61936]: DEBUG oslo.service.loopingcall [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61936) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1303.266754] env[61936]: DEBUG nova.compute.manager [-] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Deallocating network for instance {{(pid=61936) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1303.266865] env[61936]: DEBUG nova.network.neutron [-] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] deallocate_for_instance() {{(pid=61936) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1303.456436] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1303.550991] env[61936]: DEBUG nova.compute.manager [req-0fde61c0-56b4-4fe1-acc1-e310068f0d43 req-f94e5139-966b-441b-b166-62b740b0b31c service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Received event network-vif-deleted-855dd2af-6f16-4185-88e2-0022244dc30d {{(pid=61936) external_instance_event /opt/stack/nova/nova/compute/manager.py:11529}} [ 1303.551389] env[61936]: INFO nova.compute.manager [req-0fde61c0-56b4-4fe1-acc1-e310068f0d43 req-f94e5139-966b-441b-b166-62b740b0b31c service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Neutron deleted interface 855dd2af-6f16-4185-88e2-0022244dc30d; detaching it from the instance and deleting it from the info cache [ 1303.551695] env[61936]: DEBUG nova.network.neutron [req-0fde61c0-56b4-4fe1-acc1-e310068f0d43 req-f94e5139-966b-441b-b166-62b740b0b31c service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1304.027517] env[61936]: DEBUG nova.network.neutron [-] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Updating instance_info_cache with network_info: [] {{(pid=61936) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1304.054104] env[61936]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-403aca7b-655d-462e-ab35-902fcf7179d2 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.063680] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8b190f-b8a0-4931-b639-09c276899596 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.086461] env[61936]: DEBUG nova.compute.manager [req-0fde61c0-56b4-4fe1-acc1-e310068f0d43 req-f94e5139-966b-441b-b166-62b740b0b31c service nova] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Detach interface failed, port_id=855dd2af-6f16-4185-88e2-0022244dc30d, reason: Instance e6930eaf-bb63-4553-8cd2-f2eb62fca510 could not be found. {{(pid=61936) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11363}} [ 1304.450118] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1304.454756] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1304.454911] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Starting heal instance info cache {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10326}} [ 1304.455050] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Rebuilding the list of instances to heal {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10330}} [ 1304.530368] env[61936]: INFO nova.compute.manager [-] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Took 1.26 seconds to deallocate network for instance. [ 1304.959109] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] [instance: e6930eaf-bb63-4553-8cd2-f2eb62fca510] Skipping network cache update for instance because it is being deleted. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 1304.959109] env[61936]: DEBUG nova.compute.manager [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Didn't find any instances for network info cache update. {{(pid=61936) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10412}} [ 1304.959109] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1304.959109] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager.update_available_resource {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1305.036034] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1305.036465] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1305.036529] env[61936]: DEBUG nova.objects.instance [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lazy-loading 'resources' on Instance uuid e6930eaf-bb63-4553-8cd2-f2eb62fca510 {{(pid=61936) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1305.461705] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1305.569132] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618740df-9d8f-4d7b-bd6f-cb722eb3175d {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.576841] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577f6aa0-6152-48ac-a46b-07606e02cacb {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.605359] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4648257e-69ca-4c58-98ce-81885cc64ab0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.611855] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab98685-4d92-4794-bb1a-5424af792309 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.624990] env[61936]: DEBUG nova.compute.provider_tree [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1306.128522] env[61936]: DEBUG nova.scheduler.client.report [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1306.634138] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.636496] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.175s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1306.636703] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1306.636867] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61936) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1306.639374] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d040984-38d1-4e62-a6cc-9581e0ea4566 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.647250] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b9f3da-9e14-447b-ac99-2c98a98fc3cf {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.651673] env[61936]: INFO nova.scheduler.client.report [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Deleted allocations for instance e6930eaf-bb63-4553-8cd2-f2eb62fca510 [ 1306.663548] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6506bc1b-822c-404d-8ded-e2bfb61383b7 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.669846] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1fbe9e3-ef5c-48e3-bdf0-efbb684da9b9 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.701265] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180657MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=61936) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1306.701435] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1306.701630] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1307.166701] env[61936]: DEBUG oslo_concurrency.lockutils [None req-4bc7b255-8987-4846-99c1-8b90ffe7c0a5 tempest-ServerActionsTestOtherA-1117253848 tempest-ServerActionsTestOtherA-1117253848-project-member] Lock "e6930eaf-bb63-4553-8cd2-f2eb62fca510" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.516s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1307.719801] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1307.720031] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61936) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1307.732800] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ea93f3-093e-4982-851b-e7ead73e16c0 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.741254] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd82886e-a8a2-4a5d-8d5c-fe4aae6bf896 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.772608] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44364ce-e4df-4ccf-a1c7-9a35e8c21507 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.780324] env[61936]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cddf4f9-b2d2-4851-9a3f-70e58a959396 {{(pid=61936) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.793074] env[61936]: DEBUG nova.compute.provider_tree [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed in ProviderTree for provider: abd382a7-b427-4dcc-a050-3fb56fdd60ac {{(pid=61936) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1308.295571] env[61936]: DEBUG nova.scheduler.client.report [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Inventory has not changed for provider abd382a7-b427-4dcc-a050-3fb56fdd60ac based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61936) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1308.800711] env[61936]: DEBUG nova.compute.resource_tracker [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61936) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1308.800919] env[61936]: DEBUG oslo_concurrency.lockutils [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.099s {{(pid=61936) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1311.297550] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.297916] env[61936]: DEBUG oslo_service.periodic_task [None req-8bb4b9e1-496c-45fc-bf45-de1cad0130ab None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61936) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}